Menu

Blog

Archive for the ‘information science’ category: Page 60

Dec 16, 2022

Meta AI Releases Data2vec 2.0: An Efficient Self-Supervised Learning For Machine Learning Tasks

Posted by in categories: information science, robotics/AI

Self-supervised learning is a form of unsupervised learning in which the supervised learning task is constructed from raw, unlabeled data. Supervised learning is effective but usually requires a large amount of labeled data. Getting high-quality labeled data is time-consuming and resource-intensive, especially for sophisticated tasks like object detection and instance segmentation, where more in-depth annotations are sought.

Self-supervised learning aims to first learn usable representations of the data from an unlabeled pool of data by self-supervision and then to refine these representations with few labels for the supervised downstream tasks such as image classification, semantic segmentation, etc.

Self-supervised learning is at the heart of many recent advances in artificial intelligence. However, existing algorithms focus on a particular modality (such as images or text) and a high computer resource requirement. Humans, on the other hand, appear to learn significantly more efficiently than existing AI and to learn from diverse types of information consistently rather than requiring distinct learning systems for text, speech, and other modalities.

Dec 16, 2022

DALL-E 2 was the AI art generator we never knew we needed

Posted by in categories: information science, robotics/AI

DALL-E 2 transformed the world of art in 2022.

DALL-E is a system that has been around for years, but its successor, DALL-E 2, was launched this year.


Ibrahim Can/Interesting Engineering.

Continue reading “DALL-E 2 was the AI art generator we never knew we needed” »

Dec 15, 2022

Researchers develop all-optical approach to pumping chip-based nanolasers

Posted by in categories: computing, information science, internet

Researchers have developed a new all-optical method for driving multiple highly dense nanolaser arrays. The approach could enable chip-based optical communication links that process and move data faster than today’s electronic-based devices.

“The development of optical interconnects equipped with high-density nanolasers would improve information processing in the that move information across the internet,” said research team leader Myung-Ki Kim from Korea University.

“This could allow streaming of ultra-high-definition movies, enable larger-scale interactive online encounters and games, accelerate the expansion of the Internet of Things and provide the fast connectivity needed for big data analytics.”

Dec 14, 2022

Geoffrey Hinton’s Forward-Forward Algorithm Charts a New Path for Neural Networks

Posted by in categories: information science, robotics/AI

Turing Award winner and deep learning pioneer Geoffrey Hinton, one of the original proponents of backpropagation, has argued in recent years that backpropagation does not explain how the brain works. In his NeurIPS 2022 keynote speech, Hinton proposes a new approach to neural network learning: the Forward-Forward algorithm.

Dec 14, 2022

Google Labs unveils Pitchfork, an AI that can convert old code to new code and rewrite itself

Posted by in categories: information science, robotics/AI

After six decades we have finally reached controlled fusion “ignition.” Here is how it works and what it means (and doesn’t mean):

At the Lawrence Livermore National Lab (LLNL) the National Ignition Facility (NIF) starts with the Injection Laser System (ILS), a ytterbium-doped optical fiber laser (Master Oscillator) that produces a single very lower power, 1,053 nanometer (Infrared Light) beam. This single beam is split into 48 Pre-Amplifiers Modules (PAMs) that create four beams each (192 total). Each PAM conducts a two-stage amplification process via xenon flash lamps.


Self-coding and self-updating AI algorithms appear to be on the horizon. There are talks about Pitchfork AI, a top-secret Google Labs project that can independently code, refactor, and use both its own and other people’s code.

Continue reading “Google Labs unveils Pitchfork, an AI that can convert old code to new code and rewrite itself” »

Dec 14, 2022

A memristor crossbar-based learning system for scalable and energy-efficient AI

Posted by in categories: health, information science, mobile phones, robotics/AI

Deep-learning models have proven to be highly valuable tools for making predictions and solving real-world tasks that involve the analysis of data. Despite their advantages, before they are deployed in real software and devices such as cell phones, these models require extensive training in physical data centers, which can be both time and energy consuming.

Researchers at Texas A&M University, Rain Neuromorphics and Sandia National Laboratories have recently devised a new system for deep learning models more efficiently and on a larger scale. This system, introduced in a paper published in Nature Electronics, relies on the use of new training algorithms and memristor crossbar , that can carry out multiple operations at once.

“Most people associate AI with health monitoring in smart watches, face recognition in smart phones, etc., but most of AI, in terms of energy spent, entails the training of AI models to perform these tasks,” Suhas Kumar, the senior author of the study, told TechXplore.

Dec 12, 2022

What Is SEO? Reverse Engineering Google’s AI

Posted by in categories: information science, robotics/AI

Search Engine Optimization (SEO) is the process of optimizing on-page and off-page factors that impact how high a web page ranks for a specific search term. This is a multi-faceted process that includes optimizing page loading speed, generating a link building strategy, as well as learning how to reverse engineer Google’s AI by using computational thinking.

Computational thinking is an advanced type of analysis and problem-solving technique that computer programmers use when writing code and algorithms. Computational thinkers will seek the ground truth by breaking down a problem and analyzing it using first principles thinking.

Since Google does not release their secret sauce to anyone, we will rely on computational thinking. We will walk through some pivotal moments in Google’s history that shaped the algorithms that are used, and we will learn why this matters.

Dec 10, 2022

Evolutionary computation: Keith Downing at TEDxTrondheim

Posted by in categories: information science, robotics/AI, space

Keith Downing is a professor of Computer Science at the Norwegian University of Science and Technology, specializing in Artificial Intelligence and Artificial Life. He has a particular interest in evolutionary algorithms, which have applications ranging from the development of the Mars Rover antenna, patented circuits, early driverless cars, to even art. For computer scientists to learn from nature, he believes there needs to be a shift in our traditional ways thinking.

About TEDx, x = independently organized event.
In the spirit of ideas worth spreading, TEDx is a program of local, self-organized events that bring people together to share a TED-like experience. At a TEDx event, TEDTalks video and live speakers combine to spark deep discussion and connection in a small group. These local, self-organized events are branded TEDx, where x = independently organized TED event. The TED Conference provides general guidance for the TEDx program, but individual TEDx events are self-organized.* (*Subject to certain rules and regulations).

Dec 10, 2022

The Knapsack Problem & Genetic Algorithms — Computerphile

Posted by in categories: computing, genetics, information science

Tournament selection, roulette selection, mutation, crossover — all processes used in genetic algorithms. Dr Alex Turner explains using the Knapsack Problem.

https://www.facebook.com/computerphile.
https://twitter.com/computer_phile.

Continue reading “The Knapsack Problem & Genetic Algorithms — Computerphile” »

Dec 10, 2022

Why European researchers hooked up a quantum machine to a supercomputer

Posted by in categories: information science, quantum physics, supercomputing

In the cons column, quantum computers are hard to use, require a very controlled set up to operate, and have to contend with “decoherence” or losing their quantum state which gives weird results. They’re also rare, expensive, and for most tasks, way less efficient than a traditional computer.

Still, a lot of these issues can be offset by combining a quantum computer with a traditional computer, just as VTT has done. Researchers can create a hybrid algorithm that has LUMI, the traditional supercomputer, handle the parts it does best while handing off anything that could benefit from quantum computing to HELMI. LUMI can then integrate the results of HELMI’s quantum calculations, perform any additional calculations necessary or even send more calculations to HELMI, and return the complete results to the researchers.

Finland is now one of few nations in the world with a quantum computer and a supercomputer, and LUMI is the most powerful quantum-enabled supercomputer. While quantum computers are still a way from being broadly commercially viable, these kinds of integrated research programs are likely to accelerate progress. VTT is currently developing a 20-qubit quantum computer with a 50-qubit upgrade planned for 2024.

Page 60 of 280First5758596061626364Last