Toggle light / dark theme

Alex Hern reports on recent developments in artificial intelligence and how a Google employee became convinced an AI chatbot was sentient.

How to listen to podcasts: everything you need to know

Google software engineer Blake Lemoine was put on leave by his employer after claiming that the company had produced a sentient artificial intelligence and posting its thoughts online. Google said it suspended him for breaching confidentiality policies.

Lauded for its compelling action sequences and exhilarating portrayal of next-gen aerial dogfighting, Top Gun: Maverick has quickly become a monumental success at the box office. But the producers couldn’t have done it without leveraging the expertise of some of the world’s foremost experts in all things aerospace, and that includes tapping into the minds of Lockheed Martin Skunk Works engineers to craft their physics-bending Darkstar hypersonic jet.

Without wanting to give away any of the plot’s specifics, the Darkstar aircraft features early in the film as Pete “Maverick” Mitchell (played by Tom Cruise) carries out his duties as a test pilot for the US Navy. The futuristic fighter jet is a jaw-dropping introduction to the hyperreal aesthetics of the film, but may also strike a familiar chord with aviation enthusiasts due to a likeness to one of history’s most revered aerial vehicles, the SR-71 Blackbird.

When looking for some expert assistance in creating the Darkstar aircraft, the film’s producers were pointed in the direction of Lockheed Martin’s Skunk Works division, responsible for the SR-71, its forthcoming successor the SR-72 and the U-2 spy plane. This collaboration created a new outlet for expression for Skunk Works clandestine conceptual designers, in the sense that this particular aircraft design was one they could share with the world – as conceptual designer “Jim” explains in this video.

China’s first Mars mission will search for pockets of water beneath the surface that could host life.


As China’s first Mars mission, is uniquely ambitious. No nation had ever attempted to send an orbiter and rover to Mars on the first try. But China succeeded, making a historic victory.

Tianwen-1 arrived in Mars orbit as a single spacecraft. Once on Mars, the landing platform extended a ramp, allowing the Zhurong rover to roll gently onto the surface—similar to the way China’s Chang’e Moon rovers are deployed.

When did the Zhurong rover land on Mars?

In May 1997, a large earthquake shook the Kermadec Islands region in the South Pacific Ocean. A little over 20 years later, in September 2018, a second big earthquake hit the same location, its waves of seismic energy emanating from the same region.

Though the earthquakes occurred two decades apart, because they occurred in the same region, they’d be expected to send seismic waves through the Earth’s layers at the same speed, said Ying Zhou, a geoscientist with the Department of Geosciences in the Virginia Tech College of Science.

But in data recorded at four of more than 150 Global Seismographic Network stations that log seismic vibrations in real time, Zhou found an anomaly among the twin events: During the 2018 , a set of seismic waves known as SKS waves traveled about one second faster than their counterparts had in 1997.

Amazon introduced the technology at Amazon re: MARS 2022, its annual AI event centered around machine learning, automation, robotics, and space. Alexa AI head scientist Rohit Prasad referred to the upcoming feature as a way to remember friends and family members who have passed away.

“While AI can’t eliminate the pain of loss, it can definitely make their memories last,” Prasad said.

Prasad demonstrated the feature using a video of a child asking Alexa if his grandmother could finish reading him a story. In its regular Alexa voice, the smart speaker obliged; then the grandmother’s voice took over as the child flipped through his own copy of The Wizard of Oz. Though of course there’s no way for the viewer to know what the woman’s real voice actually sounds like, the grandmother’s synthesized voice admittedly sounded quite natural, speaking with the cadence of your average bedtime story reader.

Although the name and scenario are fictional, this is a question we have to confront now. In December 2021, Melbourne-based Cortical Labs grew groups of neurons (brain cells) that were incorporated into a computer chip. The resulting hybrid chip works because both brains and neurons share a common language: electricity.

In silicon computers, electrical signals travel along metal wires that link different components together. In brains, neurons communicate with each other using electric signals across synapses (junctions between nerve cells). In Cortical Labs’ Dishbrain system, neurons are grown on silicon chips. These neurons act like the wires in the system, connecting different components. The major advantage of this approach is that the neurons can change their shape, grow, replicate, or die in response to the demands of the system.

Dishbrain could learn to play the arcade game Pong faster than conventional AI systems. The developers of Dishbrain said: “Nothing like this has ever existed before … It is an entirely new mode of being. A fusion of silicon and neuron.”