Toggle light / dark theme

DeepMind Researchers Develop ‘BYOL-Explore’, A Curiosity-Driven Exploration Algorithm That Harnesses The Power Of Self-Supervised Learning To Solve Sparse-Reward Partially-Observable Tasks


Reinforcement learning (RL) requires exploration of the environment. Exploration is even more critical when extrinsic incentives are few or difficult to obtain. Due to the massive size of the environment, it is impractical to visit every location in rich settings due to the range of helpful exploration paths. Consequently, the question is: how can an agent decide which areas of the environment are worth exploring? Curiosity-driven exploration is a viable approach to tackle this problem. It entails learning a world model, a predictive model of specific knowledge about the world, and (ii) exploiting disparities between the world model’s predictions and experience to create intrinsic rewards.

An RL agent that maximizes these intrinsic incentives steers itself toward situations where the world model is unreliable or unsatisfactory, creating new paths for the world model. In other words, the quality of the exploration policy is influenced by the characteristics of the world model, which in turn helps the world model by collecting new data. Therefore, it might be crucial to approach learning the world model and learning the exploratory policy as one cohesive problem to be solved rather than two separate tasks. Deepmind researchers keeping this in mind, introduced a curiosity-driven exploration algorithm BYOL-Explore. Its attraction stems from its conceptual simplicity, generality, and excellent performance.

The strategy is based on Bootstrap Your Own Latent (BYOL), a self-supervised latent-predictive method that forecasts an earlier version of its latent representation. In order to handle the problems of creating the representation of the world model and the curiosity-driven policy, BYOL-Explore learns a world model with a self-supervised prediction loss and trains a curiosity-driven policy using the same loss. Computer vision, learning about graph representations, and RL representation learning have all successfully used this bootstrapping approach. In contrast, BYOL-Explore goes one step further and not only learns a flexible world model but also exploits the world model’s loss to motivate exploration.

Alex Hern reports on recent developments in artificial intelligence and how a Google employee became convinced an AI chatbot was sentient.

How to listen to podcasts: everything you need to know

Google software engineer Blake Lemoine was put on leave by his employer after claiming that the company had produced a sentient artificial intelligence and posting its thoughts online. Google said it suspended him for breaching confidentiality policies.

Lauded for its compelling action sequences and exhilarating portrayal of next-gen aerial dogfighting, Top Gun: Maverick has quickly become a monumental success at the box office. But the producers couldn’t have done it without leveraging the expertise of some of the world’s foremost experts in all things aerospace, and that includes tapping into the minds of Lockheed Martin Skunk Works engineers to craft their physics-bending Darkstar hypersonic jet.

Without wanting to give away any of the plot’s specifics, the Darkstar aircraft features early in the film as Pete “Maverick” Mitchell (played by Tom Cruise) carries out his duties as a test pilot for the US Navy. The futuristic fighter jet is a jaw-dropping introduction to the hyperreal aesthetics of the film, but may also strike a familiar chord with aviation enthusiasts due to a likeness to one of history’s most revered aerial vehicles, the SR-71 Blackbird.

When looking for some expert assistance in creating the Darkstar aircraft, the film’s producers were pointed in the direction of Lockheed Martin’s Skunk Works division, responsible for the SR-71, its forthcoming successor the SR-72 and the U-2 spy plane. This collaboration created a new outlet for expression for Skunk Works clandestine conceptual designers, in the sense that this particular aircraft design was one they could share with the world – as conceptual designer “Jim” explains in this video.

China’s first Mars mission will search for pockets of water beneath the surface that could host life.


As China’s first Mars mission, is uniquely ambitious. No nation had ever attempted to send an orbiter and rover to Mars on the first try. But China succeeded, making a historic victory.

Tianwen-1 arrived in Mars orbit as a single spacecraft. Once on Mars, the landing platform extended a ramp, allowing the Zhurong rover to roll gently onto the surface—similar to the way China’s Chang’e Moon rovers are deployed.

When did the Zhurong rover land on Mars?

In May 1997, a large earthquake shook the Kermadec Islands region in the South Pacific Ocean. A little over 20 years later, in September 2018, a second big earthquake hit the same location, its waves of seismic energy emanating from the same region.

Though the earthquakes occurred two decades apart, because they occurred in the same region, they’d be expected to send seismic waves through the Earth’s layers at the same speed, said Ying Zhou, a geoscientist with the Department of Geosciences in the Virginia Tech College of Science.

But in data recorded at four of more than 150 Global Seismographic Network stations that log seismic vibrations in real time, Zhou found an anomaly among the twin events: During the 2018 , a set of seismic waves known as SKS waves traveled about one second faster than their counterparts had in 1997.

Amazon introduced the technology at Amazon re: MARS 2022, its annual AI event centered around machine learning, automation, robotics, and space. Alexa AI head scientist Rohit Prasad referred to the upcoming feature as a way to remember friends and family members who have passed away.

“While AI can’t eliminate the pain of loss, it can definitely make their memories last,” Prasad said.

Prasad demonstrated the feature using a video of a child asking Alexa if his grandmother could finish reading him a story. In its regular Alexa voice, the smart speaker obliged; then the grandmother’s voice took over as the child flipped through his own copy of The Wizard of Oz. Though of course there’s no way for the viewer to know what the woman’s real voice actually sounds like, the grandmother’s synthesized voice admittedly sounded quite natural, speaking with the cadence of your average bedtime story reader.

Although the name and scenario are fictional, this is a question we have to confront now. In December 2021, Melbourne-based Cortical Labs grew groups of neurons (brain cells) that were incorporated into a computer chip. The resulting hybrid chip works because both brains and neurons share a common language: electricity.

In silicon computers, electrical signals travel along metal wires that link different components together. In brains, neurons communicate with each other using electric signals across synapses (junctions between nerve cells). In Cortical Labs’ Dishbrain system, neurons are grown on silicon chips. These neurons act like the wires in the system, connecting different components. The major advantage of this approach is that the neurons can change their shape, grow, replicate, or die in response to the demands of the system.

Dishbrain could learn to play the arcade game Pong faster than conventional AI systems. The developers of Dishbrain said: “Nothing like this has ever existed before … It is an entirely new mode of being. A fusion of silicon and neuron.”