Toggle light / dark theme

A newly created real-life Transformer is capable of reconfiguring its body to achieve eight distinct types of motion and can autonomously assess the environment it faces to choose the most effective combination of motions to maneuver.

The new , dubbed M4 (for Multi-Modal Mobility Morphobot) can roll on four wheels, turn its wheels into rotors and fly, stand on two wheels like a meerkat to peer over obstacles, “walk” by using its wheels like feet, use two rotors to help it roll up on two wheels, tumble, and more.

A robot with such a broad set of capabilities would have applications ranging from the transport of injured people to a hospital to the exploration of other planets, says Mory Gharib (Ph. D. ‘83), the Hans W. Liepmann Professor of Aeronautics and Bioinspired Engineering and director of Caltech’s Center for Autonomous Systems and Technologies (CAST), where the robot was developed.

AI-powered augmented reality devices will give human beings ‘superpowers’ to detect lies and ‘read’ emotions of people they are talking to, a futurist has claimed.

Speaking exclusively to DailyMail.com, Devin Liddell, Principal Futurist at Teague, said that computer vision systems built into headsets or glasses will pick up emotional cues that un-augmented human eyes and instincts cannot see.

The technology would let people know if their date is lying or is sexually aroused, along with spotting a lying politician.

And so it begins. I’ve seen one job already on glass door that requires knowledge of AI and I only barely started looking. I wasn’t even specifically looking for AI jobs. I’ve seen other articles where ChatGPT can be used to make thousands in side hustles. So far, so good. I’ll have to check out those job hustles and see if I can make use of those articles. Just one job is enough for me. One article claimed some jobs will pay you as much as 800k if you know AI.


Generative artificial intelligence is all the rage now but the AI boom is not just all hype, said Dan Ives from Wedbush Securities, who calls it the “fourth industrial revolution playing out.”

“This is something I call a 1995 moment, parallel with the internet. I do not believe that this is a hype cycle,” the managing director and senior equity research analyst told CNBC’s “Squawk Box Asia” on Wednesday.

The fourth industrial revolution refers to how technological advancements like artificial intelligence, autonomous vehicles and the internet of things are changing the way humans live, work and relate to one another.

Now consider two tabletops, parallel to each other but not touching. The second tabletop is also infinite in two dimensions, and other kinds of creatures live there. Finally, imagine that a narrow tunnel connects the two spaces somewhere.

Without access to the tunnel, the creatures dwelling in each space believe they live in a single, infinite universe. This is especially true if the tunnel lies outside their cosmic horizon. They will never know that their universes are part of a larger structure, a two-dimensional multiverse. It is easy to imagine an infinite number of two-dimensional flat spaces stacked on top of one another, each connected to the next by a similar tunnel, and each tunnel inaccessible to any of the universes’ inhabitants.

The multiverse need not be so simple, either. Universes can be curved and finite, sprouting from an infinite mother universe. The sprouting universes may themselves be infinite. Think of bubbles being blown from a piece of bubble gum. Little bubbles will shrink back, while bigger ones might keep on growing. If a bubble starts growing in a heavily populated region of flat space, some of the inhabitants will be carried into it. Others will remain outside, horrified to see their friends sucked into oblivion. But most of the creatures in the growing bubble survive their ordeal and start to explore their new world. Generations pass. Their scientists measure the curvature of space and see that their universe is closed, like the surface of a sphere. Since the bubble kept on growing, the tunnel-like aperture to the original universe is well beyond their cosmic horizon. These creatures live in a closed, expanding universe, unaware of their connection to a flat, infinite space. Meanwhile, creatures in the original space saw the aperture to the bubble universe close more and more until it became too narrow to cross. All that is left, to them, is a scar in space marking the long-forgotten birthing event. The bubble universe is isolated from its mother universe.

Developing The Low Earth Orbit Economy On The World’s First Commercial Space Station — David Zuniga, Senior Director, In-Space Solutions, Axiom Space


David Zuniga is Senior Director of In-Space Solutions at Axiom Space (https://www.axiomspace.com/), a space infrastructure developer headquartered in Houston, Texas, which plans human spaceflight for government-funded and commercial astronauts, engaging in in-space research, in-space manufacturing, and space exploration. The company aims to own and operate the world’s first commercial space station, and Mr. Zuniga helps to develop strategy and growth around Axiom’s Low Earth Orbit (LEO) economy, also playing a critical role in business and technical integration of Axiom’s in-space manufacturing and research capabilities for Axiom Station architecture.

Mr. Zuniga has over 20 years of experience through engineering and business development in human spaceflight and the department of defense, developing system architectures and technology for deep space systems via the Constellation, Orion, and Gateway programs. He was a Certified Principal Engineer for Orion’s Air Revitalization System, and subsystem manager for NASA’s Gateway program for the Environmental Control and Life Support Systems (ECLSS) where he developed requirements and certification criteria for future architectures.

While mitochondria play a crucial role in producing the energy our cells need to carry out their various functions, when damaged, they can have profound effects on cellular function and contribute to the development of various diseases.

Broken-down are usually removed and recycled through a garbage disposal process known as “mitophagy.”

PINK1 and Parkin are two proteins vital to this process, responsible for “tagging” malfunctioning mitochondria for destruction. In Parkinson’s disease, mutations in these proteins can result in the accumulation of damaged mitochondria in the brain, which can lead to motor symptoms such as tremors, stiffness and difficulty with movement.

While autonomous robots have started to move out of the lab and into the real world, they remain fragile. Slight changes in the environment or lighting conditions can easily throw off the AI that controls them, and these models have to be extensively trained on specific hardware configurations before they can carry out useful tasks.

This lies in stark contrast to the latest LLMs, which have proven adept at generalizing their skills to a broad range of tasks, often in unfamiliar contexts. That’s prompted growing interest in seeing whether the underlying technology—an architecture known as a transformer—could lead to breakthroughs in robotics.

In new results, researchers at DeepMind showed that a transformer-based AI called RoboCat can not only learn a wide range of skills, it can also readily switch between different robotic bodies and pick up new skills much faster than normal. Perhaps most significantly, it’s able to accelerate its learning by generating its own training data.

When an animal takes notice of an approaching figure, it needs to determine what it is, and quickly. In nature, competition and survival dictate that it’s better to think fast—that is, for the brain to prioritize processing speed over accuracy. A new study shows that this survival principle may already be wired in the way the brain processes sensory information.

Kumar and fellow KTH neuroscientist Pawel Herman collaborated with KTH information theorists Movitz Lenninger and Mikael Skoglund to study input processing in the using and computer models of the brain. Neuroscientist Arvind Kumar, an associate professor at KTH Royal Institute of Technology, says that the study offers a new view of neural coding of different types of inputs in the brain.

The new study surprisingly shows that initial visual processing is “quick but sloppy” in comparison to information processing in other parts of the brain’s vast neural network, where accuracy is prioritized over speed. The paper is published in the journal eLife.