Toggle light / dark theme

There’s a tiny, slow-burning ‘fuse’ attached to the ends of all our chromosomes, and as we naturally age, each of our cells loses more and more of that life-giving line.

Researchers in South Korea have now shown this fuse, known as the telomere, is unusually short in the cells of elderly people who are relatively healthy but have noticed early signs of depressive symptoms and cognitive decline, such as memory loss.

The randomized controlled trial presents more evidence for the telomere hypothesis of aging, which posits that all cells hit a point where they can no longer divide and replicate.

You can easily picture yourself riding a bicycle across the sky even though that’s not something that can actually happen. You can envision yourself doing something you’ve never done before—like water skiing—and maybe even imagine a better way to do it than anyone else.

Imagination involves creating a mental image of something that is not present for your senses to detect, or even something that isn’t out there in reality somewhere. Imagination is one of the key abilities that make us human. But where did it come from?

I’m a neuroscientist who studies how children acquire . I’m especially interested in the neurological mechanisms of imagination. Once we identify what brain structures and connections are necessary to mentally construct new objects and scenes, scientists like me can look back over the course of evolution to see when these emerged—and potentially gave birth to the first kinds of imagination.

Researchers from The University of Queensland have discovered the active compound from an edible mushroom that boosts nerve growth and enhances memory.

Professor Frederic Meunier from the Queensland Brain Institute said the team had identified new active compounds from the mushroom, Hericium erinaceus.

“Extracts from these so-called ‘lion’s mane’ mushrooms have been used in in Asian countries for centuries, but we wanted to scientifically determine their on ,” Professor Meunier said.

OpenAI announced its latest language model, GPT-4, but many in the AI community were disappointed by the lack of public information. Their complaints track increasing tensions in the AI world over safety.

Yesterday, OpenAI announced GPT-4, its long-awaited next-generation AI language model.


Should AI research be open or closed? Experts disagree.

Many in the AI community have criticized this decision, noting that it undermines the company’s founding ethos as a research org and makes it harder for others to replicate its work. Perhaps more significantly, some say it also makes it difficult to develop safeguards against the sort of threats posed by AI systems like GPT-4, with these complaints coming at a time of increasing tension and rapid progress in the AI world.

“I think we can call it shut on ‘Open’ AI: the 98 page paper introducing GPT-4 proudly declares that they’re disclosing *nothing* about the contents of their training set,” tweeted Ben Schmidt, VP of information design at Nomic AI, in a thread on the topic.

Earlier this month, when LinkedIn started seeding “AI-powered conversation starters” in people’s news feeds to boost engagement on its platform, the move saw more than little engagement of its own, none of it too positive.

But the truth of the matter with LinkedIn is that it’s been using a lot of AI and other kinds of automation across different aspects of its platform for years, primarily behind the scenes with how it builds and operates its network. Now, with its owner Microsoft going all-in on OpenAI, it looks like it’s becoming a more prominent part of the strategy for LinkedIn on the front end, too — with the latest coming today in the areas of LinkedIn profiles, recruitment and LinkedIn Learning.

The company is today introducing AI-powered writing suggestions, which will initially be offered to people to spruce up their LinkedIn profiles, and to recruiters writing job descriptions. Both are built on advanced GPT models, said Tomer Cohen, LinkedIn’s chief product officer. LinkedIn is using GPT-4 for personalized profiles, with GPT-3.5 for job descriptions. Alongside this, the company is also creating a bigger focus on AI in LinkedIn Learning, corralling 100 courses around the subject and adding 20 more focused just on generative AI.

As we hurtle towards a future filled with artificial intelligence, many commentators are wondering aloud whether we’re moving too fast. The tech giants, the researchers, and the investors all seem to be in a mad dash to develop the most advanced AI. But are they considering the risks, the worriers ask?

The question is not entirely moot, and rest assured that there are hundreds of incisive minds considering the dystopian possibilities — and ways to avoid them. But the fact is that the future is unknown, the implications of this powerful new technology are as unimagined as was social media at the advent of the Internet. There will be good and there will be bad, but there will be powerful artificial intelligence systems in our future and even more powerful AIs in the futures of our grandchildren. It can’t be stopped, but it can be understood.

I spoke about this new technology with Ilya Stutskeve r, a co-founder of OpenAI, the not-for-profit AI research institute whose spinoffs are likely to be among the most profitable entities on earth. My conversation with Ilya was shortly before the release of GPT-4, the latest iteration of OpenAI’s giant AI system, which has consumed billions of words of text — more than any one human could possibly read in a lifetime.

It’s part of a South Korean project to develop one of the largest semiconductor manufacturing hubs in the world.

An investment of $230 billion is expected to be made in the next two decades by Samsung to further semiconductor manufacturing.

The mega cluster, which will feature five new semiconductor plants built by Samsung, will be set in the Gyeonggi Province and is expected to be completed by 2042.


Holgs/iStock.

The investment will be a part of a South Korean project to develop one of the largest semiconductor manufacturing hubs in the world, near its capital city of Seoul. Samsung leads in the production of memory chips, smartphones, and TVs in the world.

16,000 financial advisors of the bank must be nervous.

Multinational investment management and financial services company Morgan Stanley is deploying a sophisticated chatbot to support the bank’s army of financial advisors powered by the most recent OpenAI technology, according to CNBC

The tool’s goal is to help the bank’s advisors access its data.


Getty Images.

According to Jeff McMillan, head of analytics, data, and innovation at the company’s wealth management division, the bank has tested the artificial intelligence tool with 300 advisers. It intends to make it broadly available in the upcoming months.

Called a Wolf-Rayet, these stars expel most of their outer layers into their surroundings before exploding as supernovae.

NASA’s James Webb Space Telescope has released a phenomenal image of a supernova waiting to happen. Called a Wolf-Rayet, these stars are among the most massive, luminous, and “briefly detectable” stars known. They’re at an advanced stage of stellar evolution and expel most of their outer layers into their surroundings before exploding as supernovae.

Webb had a rare sighting of a Wolf-Rayet star in June 2022. In the latest image, the telescope shows the star, WR 124, in unprecedented detail, thanks to its infrared instruments.


NASA, ESA, CSA, stsci, webb ERO production team.

Not all stars go through a brief Wolf-Rayet phase before evolving into a supernova, which is why astronomers think Webb has captured a rare phase. This particular star is 30 times the mass of the Sun and has shed 10 Suns’ worth of material – so far. According to a release, as the ejected gas moves away from the star and cools, cosmic dust forms and glows in the infrared light detectable by Webb.

Inspired by nature, these soft robots received their amphibious upgrade with the help of bistable actuators.

Researchers at Carnegie Mellon University have created a soft robot that can effortlessly transition from walking to swimming or from crawling to rolling.

“We were inspired by nature to develop a robot that can perform different tasks and adapt to its environment without adding actuators or complexity,” said Dinesh K. Patel, a postdoctoral fellow in the Morphing Matter Lab in the School of Computer Science’s Human-Computer Interaction Institute. “Our bistable actuator is simple, stable and durable, and lays the foundation for future work on dynamic, reconfigurable soft robotics.”