Toggle light / dark theme

Get the latest international news and world events from around the world.

Log in for authorized contributors

Heavy-tailed update distributions arise from information-driven self-organization in nonequilibrium learning

Like human decision-making under real-world constraints, artificial neural networks may balance free exploration in parameter space with task-relevant adaptation. In this study, we identify consistent signatures of criticality during neural network training and provide theoretical evidence that such scaling behavior arises naturally from information-driven self-organization: a dynamic balance between the maximum entropy principle that promotes unbiased exploration and mutual information constraint that relates updates with task objective. We numerically demonstrate that the power-law exponent of updates remains stable throughout training, supporting the presence of self-organized criticality.

Crazy: Scientists Compute With Human Brain Cells

Go to https://ground.news/sabine to get 40% off the Vantage plan and see through sensationalized reporting. Stay fully informed on events around the world with Ground News.

Human brains are roughly 100,000 times more energy-efficient than current AI systems. So why don’t we build computers using human brain cells? Don’t worry, researchers are one step ahead of you there – different teams across the globe are racing to develop neuron computers; processors that integrate living brain neurons into their chips. Let’s take a look at how this technology is developing and when we might see brain cells chips in the future.

Paper 1: https://www.cell.com/neuron/fulltext/.… 2: h https://www.frontiersin.org/journals/.… 👕T-shirts, mugs, posters and more: ➜ https://sabines-store.dashery.com/ 💌 Support me on Donorbox ➜ https://donorbox.org/swtg 👉 Transcript with links to references on Patreon ➜ / sabine 📝 Transcripts and written news on Substack ➜ https://sciencewtg.substack.com/ 📩 Free weekly science newsletter ➜ https://sabinehossenfelder.com/newsle… 👂 Audio only podcast ➜ https://open.spotify.com/show/0MkNfXl… 🔗 Join this channel to get access to perks ➜ / @sabinehossenfelder 📚 Buy my book ➜ https://amzn.to/3HSAWJW #science #sciencenews #tech #neuroscience.
Paper 2: h https://www.frontiersin.org/journals/.

👕T-shirts, mugs, posters and more: ➜ https://sabines-store.dashery.com/
💌 Support me on Donorbox ➜ https://donorbox.org/swtg.
👉 Transcript with links to references on Patreon ➜ / sabine.
📝 Transcripts and written news on Substack ➜ https://sciencewtg.substack.com/
📩 Free weekly science newsletter ➜ https://sabinehossenfelder.com/newsle
👂 Audio only podcast ➜ https://open.spotify.com/show/0MkNfXl
🔗 Join this channel to get access to perks ➜
/ @sabinehossenfelder.
📚 Buy my book ➜ https://amzn.to/3HSAWJW

#science #sciencenews #tech #neuroscience

Darwin Gödel Machine Explained: Self-Improving AI Agents

In this video, we dive into Darwin Gödel Machine (DGM), introduced in a recent paper from Sakana AI and the University of British Columbia.

Darwin Gödel Machine takes self-improving AI a step froward, by introducing a mechanism for an AI agent to self-improve itself.

Paper — https://arxiv.org/abs/2505.22954
Written Review — https://aipapersacademy.com/darwin-go… 🔔 Subscribe for more AI paper reviews! 📩 Join the newsletter → https://aipapersacademy.com/newsletter/ Patreon — / aipapersacademy The video was edited using VideoScribe — https://tidd.ly/44TZEiX ___________________ Chapters: 0:00 Introduction 1:54 Darwin Gödel Machine 3:59 Results.
___________________
🔔 Subscribe for more AI paper reviews!

📩 Join the newsletter → https://aipapersacademy.com/newsletter/

Patreon — / aipapersacademy.

The video was edited using VideoScribe — https://tidd.ly/44TZEiX

/* */