Toggle light / dark theme

Red meat has been a part of diets worldwide since early man. It is an excellent source of protein, vitamins (such as B vitamins) and minerals (such as iron and zinc). However, red meat has long been associated with increasing the risk of heart disease, cancer and early death. What may not be so well known is the link between red meat consumption and type 2 diabetes.

A paper published in The Lancet in September 2024 highlighted this link to type 2 diabetes using data from the Americas, the Mediterranean, Europe, south-east Asia and the Western Pacific (20 countries included).

This recent study, with nearly 2 million participants, found that high consumption of unprocessed red meat, such as beef, lamb and pork, and processed meat, such as bacon, salami and chorizo, increased the incidence of type 2 diabetes.

The tailoring of reticular materials is key for enhancing the complexity and diversity of their structure and function. Now, a series of isomeric pillar-layered metal–organic frameworks with tunable topologies have been prepared through altering the layer stacking, which enables variability on the backbone structure, pillar spatial arrangements and pore structure.

Neural network models that are able to make decisions or store memories have long captured scientists’ imaginations. In these models, a hallmark of the computation being performed by the network is the presence of stereotyped sequences of activity, akin to one-way paths. This idea was pioneered by John Hopfield, who was notably co-awarded the 2024 Nobel Prize in Physics. Whether one-way activity paths are used in the brain, however, has been unknown.

A collaborative team of researchers from Carnegie Mellon University and the University of Pittsburgh designed a clever experiment to perform a causal test of this question using a (BCI). Their findings provide empirical support of one-way activity paths in the brain and the computational principles long hypothesized by neural network models.

Stereotyped sequences of neural population activity, also known as , is believed to underlie numerous brain functions, including , sensory perception, decision making, timing, and memory, among others. The group focused on the brain’s motor system for their work, recently published in Nature Neuroscience, where neural population activity can be used to control a BCI.

In today’s AI news, OpenAI CEO Sam Altman is trying to calm the online hype surrounding his company. On Monday, the tech boss took to X to quell viral rumors that the company had achieved artificial general intelligence. “Twitter hype is out of control again,” he wrote. “We are not gonna deploy AGI next month, nor have we built it.”

In other advancements, Stuttgart, Germany-based Sereact has secured €25mn to advance its embodied AI software that enables robots to carry out tasks they were never trained to do. “With our technology, robots act situationally rather than following rigidly programmed sequences. They adapt to dynamic tasks in real-time, enabling an unprecedented level of autonomy,” said Ralf Gulde, CEO of Sereact (short for “sense, reason, act”).

Then, seven years and seven months ago, Google changed the world with the Transformer architecture, which lies at the heart of generative AI applications like OpenAI’s ChatGPT. Now Google has unveiled a new architecture called Titans, a direct evolution of the Transformer that takes us a step closer to AI that can think like humans.

And, the World Economic Forum Global Risks Report 2025 reveals a world teetering between technological triumph and profound risk. As a structural force, it “has the potential to blur boundaries between technology and humanity” and rapidly introduce novel, unpredictable challenges.

In videos, The next frontier of AI is physical AI. NVIDIA Cosmos—a platform of state-of-the-art generative world foundation models, advanced tokenizers, guardrails, and an accelerated data processing and curation pipeline—accelerates the development of physical-AI-embodied systems such as robots and autonomous vehicles.

The Big Dipper is an asterism formed by seven bright stars in the constellation Ursa Major, the Great Bear. It is one of the most recognizable star patterns in the night sky. The asterism is well-known in many cultures and goes by many other names, including the Plough, the Great Wagon, Saptarishi, and the Saucepan.

The seven stars that form the Big Dipper are: Alkaid (Eta Ursae Majoris), Mizar (Zeta Ursae Majoris), Alioth (Epsilon Ursae Majoris), Megrez (Delta Ursae Majoris), Phecda (Gamma Ursae Majoris), Dubhe (Alpha Ursae Majoris), and Merak (Beta Ursae Majoris).

In northern latitudes, the Big Dipper is visible throughout the year. It is one of the first star patterns we learn to identify, along with Orion’s Belt, Cassiopeia’s W, and the Northern Cross in Cygnus.