Toggle light / dark theme

Drones take to the waves: Saildrones are getting data where people can’t

Science fiction often paints a terrifying picture of the future—think aliens decimating humanity, à la The War of the Worlds. But sometimes the future becoming the present can be pretty amazing—who doesn’t love successful space launches majestically catapulting humans skyward?

Or take Earth’s oceans, which are currently in the middle of a technological revolution that, outside of some very nerdy circles, has gone largely unnoticed.

“We’re at the cusp of a proliferation of lots of autonomous vehicles in the ocean,” said Alex De Robertis, a biologist at the Alaska Fisheries Science Center of the National Oceanic and Atmospheric Administration (NOAA). “Things that were science fiction not so long ago are kind of routine now.” That includes saildrones, which look like oversized orange surfboards, each with a hard, carbon-fiber sail (called a wing) and a stash of scientific equipment.

NASA is creating a ChatGPT-like assistant for astronauts

Despite our intrinsic distrust of AI in space taught to us by movies like 2001: A Space Odyssey (“I’m afraid I can’t do that, Dave”), it offers large advantages to both manned and unmanned missions. To that end, NASA is developing a system that will allow astronauts to perform maneuvers, conduct experiments and more using a natural-language ChatGPT-like interface, The Guardian reported.

“The idea is to get to a point where we have conversational interactions with space vehicles and they [are] also talking back to us on alerts, interesting findings they see in the solar system and beyond,” said Dr. Larissa Suzuki, speaking at an IEEE meeting on next-gen space communication. “It’s really not like science fiction anymore.”

NASA aims to deploy the system on its Lunar Gateway, a space station that will orbit the Moon and provide support for NASA’s Artemis mission. It would use a natural language interface that allows astronauts to seek advice on experiments or conduct maneuvers without diving into complex manuals.

These Robot Arms Were Designed to Help Humans Interact With AI

A research team at the University of Tokyo is exploring the advancement of wearable robotics. Jizai Arms is a system of supernumerary robotic limbs. Up to six of these arms can be worn and controlled by the user. The limbs allow the wearer to attach, detach, replace or edit the arms. This was designed to enable social interaction between wearers to support human beings acting with robots and AIs while maintaining a sense of self-awareness and widening the possibility of human actions.

Mashable is your source for the latest in tech, culture, and entertainment.

Follow us:
Check out https://mashable.com/
Facebook: https://www.facebook.com/mashable.
Twitter: https://twitter.com/mashable.
Instagram: https://www.instagram.com/mashable

High-performance computing, AI and cognitive simulation helped LLNL conquer fusion ignition

Part 10 a series of articles describing the elements of Lawrence Livermore National Laboratory’s fusion breakthrough.

For hundreds of Lawrence Livermore National Laboratory (LLNL) scientists on the design, experimental, and modeling and simulation teams behind inertial confinement fusion (ICF) experiments at the National Ignition Facility (NIF), the results of the now-famous Dec. 5, 2022, ignition shot didn’t come as a complete surprise.

The “crystal ball” that gave them increased pre-shot confidence in a breakthrough involved a combination of detailed high-performance computing (HPC) design and a suite of methods combining physics-based simulation with machine learning. LLNL calls this “cognitive simulation,” or CogSim.

Crowdsourced Article on Artificial Intelligence

We asked YOU about your opinion!

Since we are all thinking about Artificial Intelligence (AI), we wanted to know what everyone has to say about it. We reached out using social media and collected some various ideas and opinions on the topic. Some people chose to include their names and others opted to remain anonymous. Thanks to the people who participated!

These 3 AI tools made my two-minute how-to video way more fun and engaging

This is a really cool story that shows how you can use AI. People don’t even know many AI tools that exist, like Adobe Podcast, or that you can remove the background of a video without chroma key (green screen)

Today, I’ll show you how I selected and used three AI tools to help improve a short informational video for my YouTube channel. One tool rescued the entire little project, and the others punched up the end result. For this project, we’ll be using Adobe’s Photoshop Beta, Adobe Podcast, and the latest version of Apple’s Final Cut Pro for Mac.


Generative AI isn’t just for showy and exciting projects. Here’s how I used AI tools to transform an instructional video into something much more compelling.