The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.
OpenAI has built a striking new generative video model called Sora that can take a short text description and turn it into a detailed, high-definition film clip up to a minute long.
The updated AI model can now do some seriously impressive things with long videos or text.
Google DeepMind today launched the next generation of its powerful artificial-intelligence model Gemini, which has an enhanced ability to work with large amounts of video, text, and images.
ChatGPT maker OpenAI stepped up the race in generative artificial intelligence Thursday when it unveiled its text-to-video generation tool, Sora, viewed as an impressive but potentially dangerous step in the booming AI economy amid concerns about disinformation spread.
“Game on,” said the CEO and cofounder of rival video generator Runway after OpenAI teased content from its latest AI tool.
Marco tempest is a creative technologist at the NASA jet propulsion laboratory, a director’s fellow alumni at the MIT medialab, and the founder and director of the magiclab in new york city.
Marco Tempest uses AI to bring the world’s first programmer to life.
The human brain is probably the most complex thing in the universe. Apart from the human brain, no other system can automatically acquire new information and learn new skills, perform multimodal collaborative perception and information memory processing, make effective decisions in complex environments, and work stably with low power consumption. In this way, brain-inspired research can greatly advance the development of a new generation of artificial intelligence (AI) technologies.
Powered by new machine learning algorithms, effective large-scale labeled datasets, and superior computing power, AI programs have surpassed humans in speed and accuracy on certain tasks. However, most of the existing AI systems solve practical tasks from a computational perspective, eschewing most neuroscientific details, and tending to brute force optimization and large amounts of input data, making the implemented intelligent systems only suitable for solving specific types of problems. The long-term goal of brain-inspired intelligence research is to realize a general intelligent system. The main task is to integrate the understanding of multi-scale structure of the human brain and its information processing mechanisms, and build a cognitive brain computing model that simulates the cognitive function of the brain.
The Schwartz Reisman Institute for Technology and Society and the Department of Computer Science at the University of Toronto, in collaboration with the Vector Institute for Artificial Intelligence and the Cosmic Future Initiative at the Faculty of Arts & Science, present Geoffrey Hinton on October 27, 2023, at the University of Toronto.
0:00:00 — 0:07:20 Opening remarks and introduction. 0:07:21 — 0:08:43 Overview. 0:08:44 — 0:20:08 Two different ways to do computation. 0:20:09 — 0:30:11 Do large language models really understand what they are saying? 0:30:12 — 0:49:50 The first neural net language model and how it works. 0:49:51 — 0:57:24 Will we be able to control super-intelligence once it surpasses our intelligence? 0:57:25 — 1:03:18 Does digital intelligence have subjective experience? 1:03:19 — 1:55:36 Q&A 1:55:37 — 1:58:37 Closing remarks.
Talk title: “Will digital intelligence replace biological intelligence?”
Abstract: Digital computers were designed to allow a person to tell them exactly what to do. They require high energy and precise fabrication, but in return they allow exactly the same model to be run on physically different pieces of hardware, which makes the model immortal. For computers that learn what to do, we could abandon the fundamental principle that the software should be separable from the hardware and mimic biology by using very low power analog computation that makes use of the idiosynchratic properties of a particular piece of hardware. This requires a learning algorithm that can make use of the analog properties without having a good model of those properties. Using the idiosynchratic analog properties of the hardware makes the computation mortal. When the hardware dies, so does the learned knowledge. The knowledge can be transferred to a younger analog computer by getting the younger computer to mimic the outputs of the older one but education is a slow and painful process. By contrast, digital computation makes it possible to run many copies of exactly the same model on different pieces of hardware. Thousands of identical digital agents can look at thousands of different datasets and share what they have learned very efficiently by averaging their weight changes. That is why chatbots like GPT-4 and Gemini can learn thousands of times more than any one person. Also, digital computation can use the backpropagation learning procedure which scales much better than any procedure yet found for analog hardware. This leads me to believe that large-scale digital computation is probably far better at acquiring knowledge than biological computation and may soon be much more intelligent than us. The fact that digital intelligences are immortal and did not evolve should make them less susceptible to religion and wars, but if a digital super-intelligence ever wanted to take control it is unlikely that we could stop it, so the most urgent research question in AI is how to ensure that they never want to take control.
About Geoffrey Hinton.
Geoffrey Hinton received his PhD in artificial intelligence from Edinburgh in 1978. After five years as a faculty member at Carnegie Mellon he became a fellow of the Canadian Institute for Advanced Research and moved to the Department of Computer Science at the University of Toronto, where he is now an emeritus professor. In 2013, Google acquired Hinton’s neural networks startup, DNN research, which developed out of his research at U of T. Subsequently, Hinton was a Vice President and Engineering Fellow at Google until 2023. He is a founder of the Vector Institute for Artificial Intelligence where he continues to serve as Chief Scientific Adviser.
Issues such as abrupt changes in speed limits and incomplete lane markings are among the most influential factors that can predict road crashes, finds new research by University of Massachusetts Amherst engineers. The study then used machine learning to predict which roads may be the most dangerous based on these features.
Published in the journal Transportation Research Record, the study was a collaboration between UMass Amherst civil and environmental engineers Jimi Oke, assistant professor; Eleni Christofa, associate professor; and Simos Gerasimidis, associate professor; and civil engineers from Egnatia Odos, a publicly owned engineering firm in Greece.
The most influential features included road design issues (such as changes in speed limits that are too abrupt or guardrail issues), pavement damage (cracks that stretch across the road and webbed cracking referred to as “alligator” cracking), and incomplete signage and road markings.
A research has identified and analyzed potential areas which can give a country comparative advantage and expansion in economic activities.
The findings indicates that developing countries can leverage Artificial Intelligence (AI) to achieve a faster and more sustainable growth. This has led to countries worldwide racing to harness AI to make their industries more competitive and helping to diversify economies.