Toggle light / dark theme

AI insights in a modern world with Professor Nick Bostrom, Oxford University

For decades, philosopher Nick Bostrom (director of the Future of Humanity Institute at Oxford) has led the conversation around technology and human experience (and grabbed the attention of the tech titans who are developing AI – Bill Gates, Elon Musk, and Sam Altman).

Now, a decade after his NY Times bestseller S uperintelligence warned us of what could go wrong with AI development, he flips the script in his new book Deep Utopia: Life and Meaning in a Solved World (March 27), asking us to instead consider “What could go well?”

Ronan recently spoke to Professor Nick Bostrom.

Breaking Down Barriers: Scaling Multimodal AI with CuMo

The advent of large language models (LLMs) like GPT-4 has sparked excitement around enhancing them with multimodal capabilities to understand visual data alongside text. However, previous efforts to create powerful multimodal LLMs have faced challenges in scaling up efficiently while maintaining performance. To mitigate these issues, the researchers took inspiration from the mixture-of-experts (MoE) architecture, widely used to scale up LLMs by replacing dense layers with sparse expert modules.

In the MoE approach, instead of passing inputs through a single large model, there are many smaller expert sub-models that each specialize on a subset of the data. A routing network determines which expert(s) should process each input example. It allows scaling up total model capacity in a more parameter-efficient way.

In their approach (shown in Figure 2), CuMo, the researchers integrated sparse MoE blocks into the vision encoder and the vision-language connector of a multimodal LLM. This allows different expert modules to process different parts of the visual and text inputs in parallel rather than relying on a monolithic model to analyze everything.

The connection between AI and Ancient Greek Philosophy

🤖 🏛️ Have you ever wondered about the connection between AI and Ancient Greek Philosophy?

🧔 📜 The ancient Greek philosophers, such as Aristotle, Plato, Socrates, Democritus, Epicurus and Heraclitus explored the nature of intelligence and consciousness thousands of years ago, and their ideas are still relevant today in the age of AI.

🧠 📚 Aristotle believed that there are different levels of intelligence, ranging from inanimate objects to human beings, with each level having a distinct form of intelligence. In the context of AI, this idea raises questions about the nature of machine intelligence and where it falls in the spectrum of intelligence. Meanwhile, Plato believed that knowledge is innate and can be discovered through reason and contemplation. This view has implications for AI, as it suggests that a machine could potentially have access to all knowledge, but it may not necessarily understand it in the same way that a human would.

SamuelSchmidgall/AgentClinic: Agent benchmark for medical diagnosis

From Stanford, Albert Einstein, & Johns Hopkins U: a multimodal agent benchmark to evaluate AI in simulated clinical environments.

From stanford, albert einstein, & johns hopkins U

AgentClinic: a multimodal agent benchmark to evaluate AI in simulated clinical environments abs: https://arxiv.org/abs/2405.07960 project page: https://agentclinic.github.io code: https://github.com/samuelschmidgall/agentclinic.

A new multimodal agent…


Agent benchmark for medical diagnosis. Contribute to SamuelSchmidgall/AgentClinic development by creating an account on GitHub.

AI-powered tutor Khanmigo by Khan Academy: Your 24/7 homework helper

Did you hear the news? OpenAI’s newest model can reason across audio, vision, and text in real time.

How does GPT-4o do with math tutoring? 🤔

Sal and his son test it out on a Khan Academy math problem.

You can get AI-powered math tutoring right now with Khanmigo:


AI-powered tutor Khanmigo makes homework time easy. It’s the only AI integrated into nonprofit Khan Academy’s world-class content library.

Chat GPT can now speak and sing in real time | DW News

ChatGPT 4O can now speak and sing in real time. It can even view the real world through your phone’s camera and describe what’s happening in real time.


The AI race has just shifted into high gear, with US artificial intelligence pioneers OpenAI rolling out its new interface that works with audio and vision as well as text. The new model, called GPT-4o, has gone beyond the familiar chat-bot features and is capable of real-time, near-natural voice conversations. The developer OpenAI will also make it available to free users.

ChatGPT was already able to talk to users, but with long pauses to process the data. It often seemed a bit sluggish. This was because the feature required three internal applications, the company explained: transcribing the spoken text, processing and generating, and converting the response to speech. This caused delays.

We talk to computer scientist Mike Cook from the renowned Kings College London about the new Chat GPT-4o development.

#artificialintelligence #chatgpt #openai.

/* */