Menu

Blog

Archive for the ‘robotics/AI’ category: Page 39

Mar 9, 2024

A key to the future of robots could be hiding in liquid crystals

Posted by in categories: chemistry, physics, robotics/AI

Robots and cameras of the future could be made of liquid crystals, thanks to a new discovery that significantly expands the potential of the chemicals already common in computer displays and digital watches.

The findings, a simple and inexpensive way to manipulate the molecular properties of liquid crystals with , are now published in Advanced Materials.

“Using our method, any lab with a microscope and a set of lenses can arrange the liquid crystal alignment in any pattern they’d want,” said author Alvin Modin, a doctoral researcher studying physics at Johns Hopkins. “Industrial labs and manufacturers could probably adopt the method in a day.”

Mar 9, 2024

Neuromorphic computing: The future of IoT

Posted by in categories: biological, robotics/AI

Neuromorphic computing, inspired by the intricate architecture and functionality of the human brain, represents a departure from traditional computing paradigms. Unlike conventional von Neumann architectures, which rely on sequential processing and centralized memory, neuromorphic systems emulate the parallelism, event-driven processing, and adaptive learning capabilities of biological neural networks. By leveraging principles such as massive parallelism and event-driven modality, neuromorphic computing offers a more efficient and flexible approach to processing complex data in real-time.

Advantages of Neuromorphic Computing for IoT

The adoption of neuromorphic computing in IoT promises many benefits, ranging from enhanced processing power and energy efficiency to increased reliability and adaptability. Here are some key advantages:

Mar 9, 2024

Scientists shine new light on the future of nanoelectronic devices

Posted by in categories: biotech/medical, nanotechnology, robotics/AI, solar power

Artificial intelligence (AI) has the potential to transform technologies as diverse as solar panels, in-body medical sensors and self-driving vehicles. But these applications are already pushing today’s computers to their limits when it comes to speed, memory size and energy use.

Fortunately, scientists in the fields of AI, computing and nanoscience are working to overcome these challenges, and they are using their brains as their models.

That is because the circuits, or neurons, in the have a key advantage over today’s computer circuits: they can store information and process it in the same place. This makes them exceptionally fast and energy efficient. That is why scientists are now exploring how to use materials measured in billionths of a meter— nanomaterials—to construct circuits that work like our neurons. To do so successfully, however, scientists must understand precisely what is happening within these nanomaterial circuits at the atomic level.

Mar 9, 2024

Frontiers: This paper presents a massively parallel and scalable neuromorphic cortex simulator designed for simulating large and structurally connected spiking neural networks

Posted by in categories: biological, robotics/AI

Such as complex models of various areas of the cortex. The main novelty of this work is the abstraction of a neuromorphic architecture into clusters represented by minicolumns and hypercolumns, analogously to the fundamental structural units observed in neurobiology. Without this approach, simulating large-scale fully connected networks needs prohibitively large memory to store look-up tables for point-to-point connections. Instead, we use a novel architecture, based on the structural connectivity in the neocortex, such that all the required parameters and connections can be stored in on-chip memory. The cortex simulator can be easily reconfigured for simulating different neural networks without any change in hardware structure by programming the memory. A hierarchical communication scheme allows one neuron to have a fan-out of up to 200 k neurons. As a proof-of-concept, an implementation on one Altera Stratix V FPGA was able to simulate 20 million to 2.6 billion leaky-integrate-and-fire (LIF) neurons in real time. We verified the system by emulating a simplified auditory cortex (with 100 million neurons). This cortex simulator achieved a low power dissipation of 1.62 μW per neuron. With the advent of commercially available FPGA boards, our system offers an accessible and scalable tool for the design, real-time simulation, and analysis of large-scale spiking neural networks.

Our inability to simulate neural networks in software on a scale comparable to the human brain (1011 neurons, 1014 synapses) is impeding our progress toward understanding the signal processing in large networks in the brain and toward building applications based on that understanding. A small-scale linear approximation of a large spiking neural network will not be capable of providing sufficient information about the global behavior of such highly nonlinear networks. Hence, in addition to smaller scale systems with detailed software or hardware neural models, it is necessary to develop a hardware architecture that is capable of simulating neural networks comparable to the human brain in terms of scale, with models with an intermediate level of biological detail, that can simulate these networks quickly, preferably in real time to allow interaction between the simulation and the environment.

Mar 9, 2024

Google DeepMind Introduces Two Unique Machine Learning Models, Hawk And Griffin, Combining Gated Linear Recurrences With Local Attention For Efficient Language Models

Posted by in categories: employment, robotics/AI

Artificial Intelligence (AI) and Deep Learning, with a focus on Natural Language Processing (NLP), have seen substantial changes in the last few years. The area has advanced quickly in both theoretical development and practical applications, from the early days of Recurrent Neural Networks (RNNs) to the current dominance of Transformer models.

Models that are capable of processing and producing natural language with efficiency have advanced significantly as a result of research and development in the field of neural networks, particularly with regard to managing sequences. RNN’s innate ability to process sequential data makes them well-suited for tasks involving sequences, such as time-series data, text, and speech. Though RNNs are ideally suited for these kinds of jobs, there are still problems with scalability and training complexity, particularly with lengthy sequences.

https://marktechpost-newsletter.beehiiv.com/subscribe.

Mar 9, 2024

This Machine Learning Paper Presents a General Data Generation Process for Non-Stationary Time Series Forecasting

Posted by in category: robotics/AI

One of the cornerstone challenges in machine learning, time series forecasting has made groundbreaking contributions to several domains. However, forecasting models can’t generalize the distribution shift that changes with time because time series data is inherently non-stationary. Based on the assumptions about the inter-instance and intra-instance temporal distribution shifts, two main types of techniques have been suggested to address this issue. Both stationary and nonstationary dependencies can be separated using these techniques. Existing approaches help reduce the impact of the shift in the temporal distribution. Still, they are overly prescriptive because, without known environmental labels, every sequence instance or segment might not be stable.

Before learning about the changes in the stationary and nonstationary states throughout time, there is a need to identify when the shift in the temporal distribution takes place. By assuming nonstationarity in observations, it is possible to theoretically identify the latent environments and stationary/nonstationary variables according to this understanding.

https://marktechpost-newsletter.beehiiv.com/subscribe.

Mar 9, 2024

Researchers create AI “worms” able to spread between systems — stealing private data as they go

Posted by in category: robotics/AI

Generative AI can be hijacked by self-replicating worms.

Mar 9, 2024

The Fermi Paradox: Absent Megastructures

Posted by in categories: alien life, existential risks, robotics/AI

The great mystery of where all the aliens are in our vast Universe contemplates ancient interstellar civilizations building enormous megastructures that rival worlds or even stars in the immensity… and asks why we can’t see these giant alien artifacts.

David Brin on Event Horizon with John Michael Godier: • A.I. Wars, The Fermi Paradox and Grea…
This Week in Space with Rod Pyle: • Alien Megastructures — Isaac Arthur a…

Continue reading “The Fermi Paradox: Absent Megastructures” »

Mar 9, 2024

Engineers collaborate with ChatGPT4 to design brain-inspired chips

Posted by in categories: biological, robotics/AI

Johns Hopkins electrical and computer engineers are pioneering a new approach to creating neural network chips—neuromorphic accelerators that could power energy-efficient, real-time machine intelligence for next-generation embodied systems like autonomous vehicles and robots.

Electrical and computer engineering graduate student Michael Tomlinson and undergraduate Joe Li—both members of the Andreou Lab—used natural language prompts and ChatGPT4 to produce detailed instructions to build a spiking neural network chip: one that operates much like the human brain.

Through step-by-step prompts to ChatGPT4, starting with mimicking a single biological neuron and then linking more to form a network, they generated a full that could be fabricated.

Mar 9, 2024

Conscious AI (2024) A new book available at Amazon

Posted by in category: robotics/AI

[email protected] AI: Viewing Artificial Intelligence Through the Lens of the Theory of Everything Kindle EditionDive into Conscious AI: Vi…

Page 39 of 2,165First3637383940414243Last