Menu

Blog

Page 1083

Nov 27, 2023

StreamingLLM gives language models unlimited context

Posted by in category: innovation

StreamingLLM is an innovative framework that allows large language models to handle text of infinite length without the need for finetuning. This technique preserves attention sinks to maintain a near-normal attention score distribution. When the sequence of the conversation with the LLM surpasses the model’s context length, retains the KV cache for the attention sink tokens—four initial tokens are sufficient—and discards subsequent tokens to make room for the sliding window tokens. This approach enables the model to extend its context and stabilize its performance without having to recompute the entire KV values.

“The introduction of four initial tokens, as attention sinks, suffices to restore the LLM’s performance,” the researchers write. “In contrast, adding just one or two doesn’t achieve full recovery. We believe this pattern emerges because these models didn’t include a consistent starting token across all input samples during pre-training.”

Under the framework, the KV cache comprises the attention sinks and the rolling KV cache that retains the most recent tokens vital for language modeling. The researchers emphasize the versatility of, stating, design is versatile and can be seamlessly incorporated into any autoregressive language model that employs relative positional encoding.”

Nov 27, 2023

Google Delays Release of Gemini AI That Aims to Compete With OpenAI

Posted by in categories: business, robotics/AI

Google’s company-defining effort to catch up to ChatGPT creator OpenAI is turning out to be harder than expected.

Google representatives earlier this year told some cloud customers and business partners they would get access to the company’s new conversational AI, a large language model known as Gemini, by November. But the company recently told them not to expect it until the first quarter of next year, according to two people with direct knowledge. The delay comes at a bad time for Google, whose cloud sales growth has slowed while that of its bigger rival, Microsoft, has accelerated. Part of Microsoft’s success has come from selling OpenAI’s technology to its customers.

Nov 27, 2023

Part-biological transistors change and adapt like living tissue

Posted by in categories: biological, computing

Fio Omenetto, Silklab, Tufts University.

Transistors were a pivotal innovation in the evolution of electronic technology, and they have played a critical part in the miniaturization and advancement of electronic equipment. However, they could be even better.

Nov 27, 2023

‘First Ever’ Experiments to Measure Theoretical ‘Quantum Flickering’ in an Empty Vacuum Slated for 2024

Posted by in categories: particle physics, quantum physics

German researchers hoping to be the first to successfully measure quantum flickering directly in a completely empty vacuum are setting their sights on 2024.

If successful, the first-of-their-kind experiments are expected to either confirm the existence of quantum energy in the vacuum, a core concept of quantum electrodynamics (QED), or potentially result in the discovery of previously unknown laws of nature.

Quantum Flickering, Ghost Particles, and Energy in the Vacuum.

Nov 27, 2023

Using the world’s three most powerful particle accelerators to reveal the space-time geometry of quark matter

Posted by in categories: climatology, cosmology, finance, mapping, particle physics, sustainability

Physicists from the Eötvös Loránd University (ELTE) have been conducting research on the matter constituting the atomic nucleus utilizing the world’s three most powerful particle accelerators. Their focus has been on mapping the “primordial soup” that filled the universe in the first millionth of a second following its inception.

Intriguingly, their measurements showed that the movement of observed particles bears resemblance to the search for prey of marine predators, the patterns of climate change, and the fluctuations of stock market.

In the immediate aftermath of the Big Bang, temperatures were so extreme that atomic nuclei could not exists, nor could nucleons, their building blocks. Hence, in this first instance the universe was filled with a “” of quarks and gluons.

Nov 27, 2023

New method uses crowdsourced feedback to train robots

Posted by in category: robotics/AI

To teach an AI agent a new task, like how to open a kitchen cabinet, researchers often use reinforcement learning—a trial-and-error process where the agent is rewarded for taking actions that get it closer to the goal.

In many instances, a human expert must carefully design a reward function, which is an incentive mechanism that gives the agent motivation to explore. The human expert must iteratively update that reward function as the agent explores and tries different actions. This can be time-consuming, inefficient, and difficult to scale up, especially when the is complex and involves many steps.

Researchers from MIT, Harvard University, and the University of Washington have developed a new reinforcement learning approach that doesn’t rely on an expertly designed reward function. Instead, it leverages crowdsourced , gathered from many non-expert users, to guide the agent as it learns to reach its goal. The work has been published on the pre-print server arXiv.

Nov 27, 2023

New study shows how heat can be used in computing

Posted by in categories: computing, particle physics

Physicists at Martin Luther University Halle-Wittenberg (MLU) and Central South University in China have demonstrated that, combining specific materials, heat in technical devices can be used in computing. Their discovery is based on extensive calculations and simulations. The new approach demonstrates how heat signals can be steered and amplified for use in energy-efficient data processing.

The team’s research findings have been published in the journal Advanced Electronic Materials (“PT-Symmetry Enabled Spintronic Thermal Diodes and Logic Gates.”).

Information signals are encoded as thermal spin waves (red arrows). Logical operations are realized with two magnetic strips (signal conductors) and precisely controlled with current pulses in a spacer (platinum). (Image: Berakdar group)

Nov 27, 2023

As the ISS turns 25, a look back at the space laboratory’s legacy

Posted by in categories: education, space

The ISS just celebrated its 25th anniversary — soon, the station will be hanging up its boots.

Nov 27, 2023

Deep Mind’s Student of Games AI system can beat humans at a variety of games

Posted by in categories: entertainment, robotics/AI

A team of AI researchers from EquiLibre Technologies, Sony AI, Amii and Midjourney, working with Google’s DeepMind project, has developed an AI system called Student of Games (SoG) that is capable of both beating humans at a variety of games and learning to play new ones. In their paper published in the journal Science Advances, the group describes the new system and its capabilities.

Over the past half-century, and engineers have developed the idea of machine learning and artificial intelligence, in which human-generated data is used to train computer systems. The technology has applications in a variety of scenarios, one of which is playing board and/or parlor games.

Teaching a computer to play a and then improving its capabilities to the degree that it can beat humans has become a milestone of sorts, demonstrating how far artificial intelligence has developed. In this new study, the research team has taken another step toward artificial general intelligence—in which a computer can carry out tasks deemed superhuman.

Nov 27, 2023

Everyone is talking about artificial intelligence, but we’re missing some key changes that it will unleash

Posted by in category: robotics/AI

As an optimist, I believe it will be a catalyst for changes that will help all of us to learn faster and achieve more of our potential.

Focus on the data

Where do I start? Let me start by noting that, in all the conversations about artificial intelligence, very few people are talking about the data. Most people don’t recognize that AI is actually extremely stupid without data. Data is the fuel that shapes the intelligence of AI. Everyone seems to assume that more and more data will be available as AI evolves. But is that assumption valid?