Toggle light / dark theme

New physics-based self-learning machines could replace current artificial neural networks and save energy

Artificial intelligence not only affords impressive performance, but also creates significant demand for energy. The more demanding the tasks for which it is trained, the more energy it consumes.

Víctor López-Pastor and Florian Marquardt, two scientists at the Max Planck Institute for the Science of Light in Erlangen, Germany, present a method by which could be trained much more efficiently. Their approach relies on instead of the digital currently used. The work is published in the journal Physical Review X.

The amount of energy required to train GPT-3, which makes ChatGPT an eloquent and apparently well-informed Chatbot, has not been revealed by Open AI, the company behind that artificial intelligence (AI). According to the German statistics company Statista, this would require 1,000 megawatt hours—about as much as 200 German households with three or more people consume annually. While this has allowed GPT-3 to learn whether the word “deep” is more likely to be followed by the word “sea” or “learning” in its , by all accounts it has not understood the underlying meaning of such phrases.

Generative AI’s Biggest Security Flaw Is Not Easy to Fix

It’s easy to trick the large language models powering chatbots like OpenAI’s ChatGPT and Google’s Bard. In one experiment in February, security researchers forced Microsoft’s Bing chatbot to behave like a scammer. Hidden instructions on a web page the researchers created told the chatbot to ask the person using it to hand over their bank account details. This kind of attack, where concealed information can make the AI system behave in unintended ways, is just the beginning.

Hundreds of examples of “indirect prompt injection” attacks have been created since then. This type of attack is now considered one of the most concerning ways that language models could be abused by hackers. As generative AI systems are put to work by big corporations and smaller startups, the cybersecurity industry is scrambling to raise awareness of the potential dangers. In doing so, they hope to keep data—both personal and corporate—safe from attack. Right now there isn’t one magic fix, but common security practices can reduce the risks.

“Indirect prompt injection is definitely a concern for us,” says Vijay Bolina, the chief information security officer at Google’s DeepMind artificial intelligence unit, who says Google has multiple projects ongoing to understand how AI can be attacked. In the past, Bolina says, prompt injection was considered “problematic,” but things have accelerated since people started connecting large language models (LLMs) to the internet and plug-ins, which can add new data to the systems. As more companies use LLMs, potentially feeding them more personal and corporate data, things are going to get messy. “We definitely think this is a risk, and it actually limits the potential uses of LLMs for us as an industry,” Bolina says.

Futuristic robotaxi could be coming soon, claims Cruise

Robotaxi company Cruise is “just days away” from getting regulatory approval that would pave the way for mass production of its purpose-built driverless vehicle, CEO Kyle Vogt said on Thursday in comments reported by the Detroit Free Press.

General Motors-backed Cruise unveiled the vehicle — called Origin — in early 2020, presenting the kind of driverless car that we all dreamed of when R&D in the sector kicked off years ago; a vehicle without a steering wheel and without pedals. A vehicle with passenger seats only.

AI Research Lab Imbue Nabs $200 Million For Speculative Bet To Build AI ‘Agents’

The startup, one of very few woman-led AI unicorns, has a $1 billion valuation and access to 10,000 Nvidia H100 GPUs, but its founders say it could be years away from revealing a product.

It was at a San Francisco party hosted by Greg Brockman nearly a decade ago — around the time the Stripe CTO would leave to co-found an AI research lab called OpenAI — that entrepreneurs Kanjun Qiu and Josh Albrecht met a crypto mogul named Jed McCaleb.

As Qiu and Albrecht launched two now-defunct startups, and McCaleb became a regular at The Archive, their group house for founders and AI researchers near Dolores Park, they talked periodically about launching their own AI lab. Several of their housemates had… More.


Imbue has a $1 billion valuation and access to 10,000 Nvidia H100 GPUs, but its founders Kanjun Qiu and Josh Albrecht say it could be years away from revealing a product.

This USC Professor is Developing an Artificial Brain

Alice Parker is working on how to mimic disorders such as schizophrenia.

People have expected great things from Alice Parker, who was raised in a family of distinguished scientists and engineers. And Parker, emerita professor of electrical and computer engineering at the University of Southern California has delivered. She helped develop high-level (behavioral) synthesis, an automated computer design process that assists with the transformation of a behavioral description of hardware into a model of its logic and memory circuits.

Her father, a chemist, was on the team that first synthesized vitamin B1 at pharmaceutical company Merck in New Jersey. In 1941 her uncle Edward Wenk Jr., was… More.

The Secret to Nvidia’s Success

Chief scientist Bill Dally explains the 4 ingredients that brought Nvidia so far.

Nvidia is riding high at the moment. The company has managed to increase the performance of its chips on AI tasks a thousandfold over the past 10 years, it’s raking in money, and it’s reportedly very hard to get your hands on its newest AI-accelerating GPU, the H100.

How did Nvidia get here? The company’s chief scientist, Bill Dally, managed to sum it all up in a single slide during his keynote address to the IEEE’s Hot Chips 2023 symposium in Silicon Valley on high-performance microprocessors last week. Moore’s Law was a surprisingly small part of Nvidia’s magic and new number formats a very large part. Put it… More.

ChatGPT Makes OK Clinical Decisions—Usually

But don’t think about replacing your doctor with a chatbot now, or ever.

Could ChatGPT someday assist doctors in diagnosing patients? It might one day be possible. In a recent study, researchers fed ChatGPT information from fictional patients found in a online medical reference manual to find out how well the chatbot could make clinical decisions such as diagnosing patients and prescribing treatments. The researchers found that ChatGPT was 72 percent accurate in its decisions, although the bot was better at some kinds of clinical tasks than others. It also showed no evidence of bias based on age or gender. Though the study was small and did not use real patient data, the findings point to the… More.

China Joins Large Language Model Craze

This post is also available in: he עברית (Hebrew)

China has already released over 70 artificial intelligence large language models (LLMs), with more applications being filed every day.

Robin Li, CEO of Baidu said at an industry event in Beijing that more than 70 LLMs have been released in China, which include chatbots from the facial recognition company SenseTime and AI startups Baichuan Intelligent Technology, Zhipu AI, and MiniMax.

Researchers Experiment with Flamingo & Dall-E; the Results Will Surprise You

While text-based AI models have been found coordinating amongst themselves and developing a language of their own, communication between image-based models remained an unexplored territory, until now. A group of researchers set out to find how well Google Deepmind’s Flamingo and OpenAI’s Dall-E understand each other — their synergy is impressive.

Despite the closeness of the image captioning and text-to-image generation tasks, they are often studied in isolation from each other, i.e the information exchange between these models remains a question someone never looked for an answer to. Researchers from LMU Munich, Siemens AG, and the University of Oxford wrote a paper titled, ‘Do Flamingo and DALL-E Understand Each Other?‘investigating the communication between image captioning and text-to-image models.

The team proposes a reconstruction task where Flamingo generates a description for a given image and DALL-E uses this description as input to synthesise a new image. They argue that these models understand each other if the generated image is similar to the given image. Specifically, they studied the relationship between the quality of the image reconstruction and that of the text generation. As a result, they found that a better caption is the one that leads to better visuals and vice-versa.