Toggle light / dark theme

Creating realistic 3D models for applications like virtual reality, filmmaking, and engineering design can be a cumbersome process requiring lots of manual trial and error.

While generative artificial intelligence models for images can streamline artistic processes by enabling creators to produce lifelike 2D images from text prompts, these models are not designed to generate 3D shapes. To bridge the gap, a recently developed technique called Score Distillation leverages 2D image generation models to create 3D shapes, but its output often ends up blurry or cartoonish.

MIT researchers explored the relationships and differences between the algorithms used to generate 2D images and 3D shapes, identifying the root cause of lower-quality 3D models. From there, they crafted a simple fix to Score Distillation, which enables the generation of sharp, high-quality 3D shapes that are closer in quality to the best model-generated 2D images.


The dream of many – to try the taste through a monitor – is getting closer.

Remove ads and unlock PREMIUM content

A team of biomedical engineers and virtual reality experts has developed a groundbreaking lollipop-shaped interface that simulates taste in virtual reality.

The Matrix is a groundbreaking AI model capable of generating infinite, high-quality video worlds in real time, offering unmatched interactivity and adaptability. Developed using advanced techniques like the Video Diffusion Transformer and Swin-DPM, it enables seamless, frame-level precision for creating dynamic, responsive simulations. This innovation surpasses traditional systems, making it a game-changer for gaming, autonomous vehicle testing, and virtual environments.

🔍 Key Topics Covered:
The Matrix AI model and its ability to generate infinite, interactive video worlds.
Real-time applications in gaming, autonomous simulations, and dynamic virtual environments.
Revolutionary AI techniques like Video Diffusion Transformer, Swin-DPM, and Interactive Modules.

🎥 What You’ll Learn:
How The Matrix AI redefines video generation with infinite-length, high-quality simulations.
The transformative impact of real-time interactivity and domain generalization in AI-driven worlds.
Why this breakthrough is a game-changer for industries like gaming, VR, and autonomous systems.

📊 Why This Matters:

Researchers from Seoul National University College of Engineering announced they have developed an optical design technology that dramatically reduces the volume of cameras with a folded lens system utilizing “metasurfaces,” a next-generation nano-optical device.

By arranging metasurfaces on the so that light can be reflected and moved around in the glass substrate in a folded manner, the researchers have realized a with a thickness of 0.7mm, which is much thinner than existing refractive lens systems. The research was published on Oct. 30 in the journal Science Advances.

Traditional cameras are designed to stack multiple glass lenses to refract light when capturing images. While this structure provided excellent high-quality images, the thickness of each lens and the wide spacing between lenses increased the overall bulk of the camera, making it difficult to apply to devices that require ultra-compact cameras, such as virtual and augmented reality (VR-AR) devices, smartphones, endoscopes, drones, and more.

A paper published in Proceedings of the 37th Annual ACM Symposium on User Interface Software and Technology, by researchers in Carnegie Mellon University’s Human-Computer Interaction Institute, introduces EgoTouch, a tool that uses artificial intelligence to control AR/VR interfaces by touching the skin with a finger.

It allows multiple users to walk in any direction without colliding, enhancing VR immersion. Developed by Disney Imagineer Lanny Smoot, this innovation could revolutionize VR experiences and stage performances. (Video Credit: Disney Parks/YouTube)

Modern imaging systems, such as those used in smartphones, virtual reality (VR), and augmented reality (AR) devices, are constantly evolving to become more compact, efficient, and high-performing. Traditional optical systems rely on bulky glass lenses, which have limitations like chromatic aberrations, low efficiency at multiple wavelengths, and large physical sizes. These drawbacks present challenges when designing smaller, lighter systems that still produce high-quality images.

MIT CSAIL researchers have developed a generative AI system, LucidSim, to train robots in virtual environments for real-world navigation. Using ChatGPT and physics simulators, robots learn to traverse complex terrains. This method outperforms traditional training, suggesting a new direction for robotic training.


A team of roboticists and engineers at MIT CSAIL, Institute for AI and Fundamental Interactions, has developed a generative AI approach to teaching robots how to traverse terrain and move around objects in the real world.

The group has published a paper describing their work and possible uses for it on the arXiv preprint server. They also presented their ideas at the recent Conference on Robot Learning (CORL 2024), held in Munich Nov. 6–9.

Artificial Intelligence is everywhere in Europe.

While some are worried about its long-term impact, a team of researchers at the University of Technology in Vienna is working on responsible ways to use AI.

Watch more 👉


From industry to healthcare to the media and even the creative arts, artificial intelligence is already having an impact on our daily lives. It’s hailed by advocates as a gift to humanity, but others worry about the long-term effects on society.