Toggle light / dark theme

AI researchers expose critical vulnerabilities within major large language models

Large Language Models (LLMs) such as ChatGPT and Bard have taken the world by storm this year, with companies investing millions to develop these AI tools, and some leading AI chatbots being valued in the billions.

These LLMs, which are increasingly used within AI chatbots, scrape the entire Internet of information to learn and to inform answers that they provide to user-specified requests, known as “prompts.”

However, computer scientists from the AI security start-up Mindgard and Lancaster University in the UK have demonstrated that chunks of these LLMs can be copied in less than a week for as little as $50, and the information gained can be used to launch targeted attacks.

First supernova detected, confirmed, classified and shared by AI

A fully automated process, including a brand-new artificial intelligence (AI) tool, has successfully detected, identified and classified its first supernova.

Developed by an led by Northwestern University, the new system automates the entire search for new supernovae across the night sky—effectively removing humans from the process. Not only does this rapidly accelerate the process of analyzing and classifying new supernova candidates, it also bypasses .

The team alerted the astronomical community to the launch and success of the new tool, called the Bright Transient Survey Bot (BTSbot), this week. In the past six years, humans have spent an estimated total of 2,200 hours visually inspecting and classifying supernova candidates. With the new tool now officially online, researchers can redirect this precious time toward other responsibilities in order to accelerate the pace of discovery.

Could AI communicate with aliens better than we could?

Consider the potential problems. Number one would be that any potential aliens we encounter won’t be speaking a human language. Number two would be the lack of knowledge about the aliens’ culture or sociology — even if we could translate, we might not understand what relevance it has to their cultural touchstones.

Eamonn Kerins, an astrophysicist from the Jodrell Bank Centre for Astrophysics at the University of Manchester in the U.K., thinks that the aliens themselves might recognize these limitations and opt to do some of the heavy lifting for us by making their message as simple as possible.

“One might hope that aliens who want to establish contact might be attempting to make their signal as universally understandable as possible,” said Kerins in a Zoom interview. “Maybe it’s something as basic as a mathematical sequence, and already that conveys the one message that perhaps they hoped to send in the first place, which is that we’re here, you’re not alone.”

Elon Musk Unveils X.ai: A Game-Changer For Investors

The richest man in the world is building a super-intelligent AI to understand the true nature of the universe. This is what the project means for investors.

Elon Musk held a Twitter Spaces event in early July to reveal X.ai, his newest AI business. X.ai researchers will focus on science, while also building applications for enterprises and consumers.

To participate, investors should continue to buy Arista Networks ANET (ANET).

Visual Question Answering with Frozen Large Language Models

In this article we’ll use a Q-Former, a technique for bridging computer vision and natural language models, to create a visual question answering system. We’ll go over the necessary theory, following the BLIP-2 paper, then implement a system which can be used to talk with a large language model about an image.

Who is this useful for? Data scientists interested in computer vision, natural language processing, and multimodal modeling.

How advanced is this post? Intermediate. You might struggle if you don’t have some experience in both computer vision and natural language processing.

Multimodality and Large Multimodal Models (LMMs)

For a long time, each ML model operated in one data mode – text (translation, language modeling), image (object detection, image classification), or audio (speech recognition).

However, natural intelligence is not limited to just a single modality. Humans can read and write text. We can see images and watch videos. We listen to music to relax and watch out for strange noises to detect danger. Being able to work with multimodal data is essential for us or any AI to operate in the real world.

OpenAI noted in their GPT-4V system card that “incorporating additional modalities (such as image inputs) into LLMs is viewed by some as a key frontier in AI research and development.”