Menu

Blog

Archive for the ‘existential risks’ category: Page 24

May 30, 2023

Top AI researchers and CEOs warn against ‘risk of extinction’ in 22-word statement

Posted by in categories: biotech/medical, existential risks, robotics/AI

It’s another high-profile warning about AI risk that will divide experts. Signatories include Google DeepMind CEO Demis Hassabis and OpenAI CEO Sam Altman.

A group of top AI researchers, engineers, and CEOs have issued a new warning about the existential threat they believe that AI poses to humanity.

The 22-word statement, trimmed short to make it as broadly acceptable as possible, reads as follows: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”

Continue reading “Top AI researchers and CEOs warn against ‘risk of extinction’ in 22-word statement” »

May 30, 2023

We Are (Probably) Safe From Asteroids For 1,000 Years, Say Scientists

Posted by in categories: asteroid/comet impacts, existential risks

When will an asteroid hit Earth and wipe us out? Not for at least 1,000 years, according to a team of astronomers. Probably.

Either way, you should get to know an asteroid called 7482 (1994 PC1), the only one known whose orbital path will cross that of Earth’s consistently for the next millennium—and thus has the largest probability of a “deep close encounter” with us, specifically in 502 years. Possibly.

Published on a preprint archive and accepted for publication in The Astronomical Journal, the paper states that astronomers have almost found all the kilometer-sized asteroids. There’s a little under 1,000 of them.

May 24, 2023

Whole Brain Emulation

Posted by in categories: existential risks, mapping, neuroscience, robotics/AI

I had an amazing experience at the Foresight Institute’s Whole-Brain Emulation (WBE) Workshop at a venue near Oxford! For more information and a list of participants, see: https://foresight.org/whole-brain-emulation-workshop-2023/ I had the opportunity to work within a group of some of the most brilliant, ambitious, and visionary people I’ve ever encountered on the quest for recreating the human brain in a computer. We also discussed in depth the existential risks of upcoming artificial superintelligence and how to mitigate these risks, perhaps with the aid of WBE.

My subgroup focused on exploring the challenge of human connectomics (mapping all of the neurons and synapses in the brain).


WBE is a potential technology to generate software intelligence that is human-aligned simply by being based directly on human brains. Generally past discussions have assumed a fairly long timeline to WBE, while past AGI timelines had broad uncertainty. There were also concerns that the neuroscience of WBE might boost AGI capability development without helping safety, although no consensus did develop. Recently many people have updated their AGI timelines towards earlier development, raising safety concerns. That has led some people to consider whether WBE development could be significantly speeded up, producing a differential technology development re-ordering of technology arrival that might lessen the risk of unaligned AGI by the presence of aligned software intelligence.

May 23, 2023

Artificial Intelligence Explosion: How AI May Cause The End of The World??

Posted by in categories: existential risks, robotics/AI

Artificial intelligence is a superior lifeform that humans are creating, and many AI researchers have outlined various scenarios in which this technology can pose an existential risk to humanity that could result in the literal end of the world.

Deep Learning AI Specialization: https://imp.i384100.net/GET-STARTED
AI Marketplace: https://taimine.com/

Continue reading “Artificial Intelligence Explosion: How AI May Cause The End of The World??” »

May 13, 2023

Advanced Life Should Have Already Peaked Billions of Years Ago

Posted by in categories: alien life, existential risks, information science

Did humanity miss the party? Are SETI, the Drake Equation, and the Fermi Paradox all just artifacts of our ignorance about Advanced Life in the Universe? And if we are wrong, how would we know?

A new study focusing on black holes and their powerful effect on star formation suggests that we, as advanced life, might be relics from a bygone age in the Universe.

Universe Today readers are familiar with SETI, the Drake Equation, and the Fermi Paradox. All three are different ways that humanity grapples with its situation. They’re all related to the Great Question: Are We Alone? We ask these questions as if humanity woke up on this planet, looked around the neighbourhood, and wondered where everyone else was. Which is kind of what has happened.

Apr 30, 2023

The ‘Don’t Look Up’ Thinking That Could Doom Us With AI

Posted by in categories: asteroid/comet impacts, existential risks, robotics/AI

Suppose a large inbound asteroid were discovered, and we learned that half of all astronomers gave it at least 10% chance of causing human extinction, just as a similar asteroid exterminated the dinosaurs about 66 million years ago. Since we have such a long history of thinking about this threat and what to do about it, from scientific conferences to Hollywood blockbusters, you might expect humanity to shift into high gear with a deflection mission to steer it in a safer direction.

Sadly, I now feel that we’re living the movie Don’t look up for another existential threat: unaligned superintelligence. We may soon have to share our planet with more intelligent “minds” that care less about us than we cared about mammoths. A recent survey showed that half of AI researchers give AI at least 10% chance of causing human extinction. Since we have such a long history of thinking about this threat and what to do about it, from scientific conferences to Hollywood blockbusters, you might expect that humanity would shift into high gear with a mission to steer AI in a safer direction than out-of-control superintelligence. Think again: instead, the most influential responses have been a combination of denial, mockery, and resignation so darkly comical that it’s deserving of an Oscar.

Continue reading “The ‘Don’t Look Up’ Thinking That Could Doom Us With AI” »

Apr 29, 2023

We Solved The Fermi Paradox

Posted by in category: existential risks

It’s possibly the most famous question in all of science — where is everyone? Join us today for deep dive into Fermi Paradox.

Thanks to NordVPN for sponsoring this episode. 🌏 Get exclusive NordVPN deal here ➵ https://NordVPN.com/coolworlds It’s risk free with Nord’s 30 day money-back guarantee!✌

Continue reading “We Solved The Fermi Paradox” »

Apr 28, 2023

Huge cache of mammal genomes offers fresh insights on human evolution

Posted by in categories: asteroid/comet impacts, biotech/medical, evolution, existential risks, genetics

Using Zoonomia’s data, researchers have also constructed a phylogenetic tree that estimates when each mammalian species diverged from its ancestors5. This analysis lends support to the hypothesis that mammals had already started evolutionarily diverging before Earth was struck by the asteroid that killed the dinosaurs about 65 million years ago — but that they diverged much more rapidly afterwards.

Only the beginning

The Zoonomia Project is just one of dozens of efforts to sequence animal genomes. Another large effort is the Vertebrate Genomes Project (VGP), which aims to generate genomes for roughly all 71,000 living vertebrate species, which include mammals, reptiles, fish, birds and amphibians. Although the two projects are independent of one another, many researchers are a part of both, says Haussler, who is a trustee of the VGP.

Apr 28, 2023

There Is No A.I.

Posted by in categories: existential risks, robotics/AI

“As a computer scientist, I don’t like the term ” A.I.” In fact, I think it’s misleading—maybe even a little dangerous. Everybody’s already using the term, and it might seem a little late in the day to be arguing about it. But we’re at the beginning of a new technological era—and the easiest way to mismanage a technology is to misunderstand it.

The term artificial intelligence has a long history—it was coined in the nineteen-fifties, in the early days of computers. More recently, computer scientists have grown up on movies like The Terminator and The Matrix, and on characters like Commander Data, from Star Trek: The Next Generation. These cultural touchstones have become an almost religious mythology in tech culture. It’s only natural that computer scientists long to create A.I. and realize a long-held dream.

What’s striking, though, is that many of the people who are pursuing the A.I. dream also worry that it might mean doomsday for mankind. It is widely stated, even by scientists at the very center of today’s efforts, that what A.I. researchers are doing could result in the annihilation of our species, or at least in great harm to humanity, and soon. In a recent poll, half of A.I. scientists agreed that there was at least a ten-per-cent chance that the human race would be destroyed by A.I.

Apr 27, 2023

Genomes from 240 mammalian species reveal what makes the human genome unique

Posted by in categories: biotech/medical, evolution, existential risks, genetics, health

Over the past 100 million years, mammals have adapted to nearly every environment on Earth. Scientists with the Zoonomia Project have been cataloging the diversity in mammalian genomes by comparing DNA sequences from 240 species that exist today, from the aardvark and the African savanna elephant to the yellow-spotted rock hyrax and the zebu.

This week, in several papers in a special issue of Science, the Zoonomia team has demonstrated how can not only shed light on how certain species achieve extraordinary feats, but also help scientists better understand the parts of our genome that are functional and how they might influence health and disease.

In the new studies, the researchers identified regions of the genomes, sometimes just single letters of DNA, that are most conserved, or unchanged, across mammalian species and millions of years of evolution—regions that are likely biologically important. They also found part of the genetic basis for uncommon mammalian traits such as the ability to hibernate or sniff out faint scents from miles away. And they pinpointed species that may be particularly susceptible to extinction, as well as genetic variants that are more likely to play causal roles in rare and common human diseases.

Page 24 of 149First2122232425262728Last