Toggle light / dark theme

Cool Qubits Make Faster Decisions

Classical machine learning has benefited several physics subfields, from materials science to medical imaging. Implementing machine-learning algorithms on quantum computers could expand their use to more complex problems and to datasets that are inherently quantum. Nayeli Rodríguez-Briones at the Technical University of Vienna and Daniel Park at Yonsei University in South Korea have now proposed a thermodynamics-inspired protocol that could make quantum machine-learning techniques more efficient [1].

In one common classical machine-learning task, a system is trained on a known dataset and then challenged to classify new data. Its output quantifies both the classification and that classification’s uncertainty. Once the system’s parameters are fixed, evaluating the same data yields the same output. In contrast, the output of a quantum machine-learning algorithm is read out as binary measurements of qubits, which are inherently probabilistic. Because a single measurement provides only limited information, the computation must be repeated many times.

Rodríguez-Briones and Park recognized that how clearly a quantum computer reveals its output is determined by entropy. When the readout qubit is highly polarized—strongly favoring one outcome—its entropy is low. Few repetitions are needed to obtain a firm result. An unpolarized, high-entropy readout qubit returns both states more evenly, meaning more repetitions are required. The researchers showed that the readout qubit’s polarity can be increased by transferring its entropy to ancillary qubits, effectively cooling one while warming the others. Between runs, the ancillary qubits are reset by coupling them to a heat bath. Crucially, this entropy transfer affects the readout qubit’s degree of polarization without changing the encoded decision. The upshot: A given result can be arrived at with fewer repetitions.

Seeing global trade through the lens of physics

New research from the Complexity Science Hub (CSH) shows why widely used algorithms for measuring economic complexity produce trustworthy results and how these tools may benefit diverse areas such as ecology, social science, and agentic AI. The paper is published in the journal Physical Review E.

AI-generated Slopoly malware used in Interlock ransomware attack

A new malware strain dubbed Slopoly, likely created using generative AI tools, allowed a threat actor to remain on a compromised server for more than a week and steal data in an Interlock ransomware attack.

The breach started with a ClickFix ruse, and in later stages of the attack, the hackers deployed the Slopoly backdoor as a PowerShell script acting as a client for the command-and-control (C2) framework.

IBM X-Force researchers analyzed the script and found strong indicators that it was created using a large language model (LLM), but could not determine which one.

Rates of Unbalanced Chromosome Rearrangements Associated with Pericentric and Paracentric Inversions: Analysis of Molecular Chromosome Results in Embryo Samples

PGT-SR reveals that even small pericentric and paracentric inversions carry a small but measurable reproductive risk, challenging assumptions of minimal impact in IVF outcomes.


This website uses a security service to protect against malicious bots. This page is displayed while the website verifies you are not a bot.

AI is homogenizing human expression and thought, computer scientists and psychologists say

AI chatbots are standardizing how people speak, write, and think. If this homogenization continues unchecked, it risks reducing humanity’s collective wisdom and ability to adapt, computer scientists and psychologists argue in an opinion paper published in Trends in Cognitive Sciences.

They say that AI developers should incorporate more real-world diversity into large language model (LLM) training sets, not only to help preserve human cognitive diversity, but also to improve chatbots’ reasoning abilities.

What Happens When AI Runs the Entire Economy?

What happens when AI controls prices, jobs, markets, and growth itself? Explore the future of an economy run by machines—and what it means for work, power, and humanity.

Get Nebula using my link for 50% off an annual subscription: https://go.nebula.tv/isaacarthur Watch my exclusive video Lazarus Protocols: https://nebula.tv/videos/isaacarthur-lazarus-protocols-reviv…extinction Check out Abolish Everything https://nebula.tv/abolish?ref=isaacarthur.

🛒 SFIA Merchandise: https://isaac-arthur-shop.fourthwall.com/ 🌐 Visit our Website: http://www.isaacarthur.net ❤️ Support us on Patreon: https://www.patreon.com/IsaacArthur ⭐ Support us on Subscribestar: https://www.subscribestar.com/isaac-arthur 👥 Facebook Group: https://www.facebook.com/groups/1583992725237264/ 📣 Reddit Community: https://www.reddit.com/r/IsaacArthur/ 🐦 Follow on Twitter / X: https://twitter.com/Isaac_A_Arthur 💬 SFIA Discord Server: https://discord.gg/53GAShE Credits: What Happens When AI Runs the Entire Economy? Written, Produced & Narrated by: Isaac Arthur Select imagery/video supplied by Getty Images.

Chapters 0:00 Intro — The Invisible Hand Becomes a Neural Network 2:58 What Does “Running the Economy” Actually Mean? 5:30 What Does “Running the Economy” Actually Mean? 10:19 Labor in an AI-Run Economy 14:41 Who Programs the Economy’s Values? 17:01 Government, Power, and Economic Sovereignty 20:21 So, Can Humans Stay in the Loop? 22:56 The Best-Case and Worst-Case Futures 24:53 Abolish Everything 25:57 The Last Economic Decisions We Ever Make.

Read more

Human brain and AI speech recognition decode speech in similar step-by-step stages, study finds

Over the past decades, computer scientists have developed numerous artificial intelligence (AI) systems that can process human speech in different languages. The extent to which these models replicate the brain processes via which humans understand spoken language, however, has not yet been clearly determined.

Researchers at Columbia University, IBM Research and the Feinstein Institutes for Medical Research recently carried out a study aimed at comparing how automatic speech recognition (ASR) systems and the human brain decode speech. Their findings, published in Nature Machine Intelligence, suggest that activity in specific brain regions while people make sense of spoken language corresponds to specific stages in the processing of speech by AI models.

“The core mystery we wanted to solve is how the human brain performs the incredible computational feat of turning raw acoustic vibrations, the sounds of speech, into discrete linguistic meaning,” Nima Mesgarani, senior author of the paper, told Tech Xplore. “We now have AI systems that match human performance in transcribing speech, but we didn’t know if they were reaching those solutions independently or if they had converged on the same strategy as our biology.”

/* */