Menu

Blog

Page 4298

Apr 5, 2022

Pathways Language Model (PaLM): Scaling to 540 Billion Parameters for Breakthrough Performance

Posted by in categories: innovation, robotics/AI

In recent years, large neural networks trained for language understanding and generation have achieved impressive results across a wide range of tasks. GPT-3 first showed that large language models (LLMs) can be used for few-shot learning and can achieve impressive results without large-scale task-specific data collection or model parameter updating. More recent LLMs, such as GLaM, LaMDA, Gopher, and Megatron-Turing NLG, achieved state-of-the-art few-shot results on many tasks by scaling model size, using sparsely activated modules, and training on larger datasets from more diverse sources. Yet much work remains in understanding the capabilities that emerge with few-shot learning as we push the limits of model scale.

Last year Google Research announced our vision for Pathways, a single model that could generalize across domains and tasks while being highly efficient. An important milestone toward realizing this vision was to develop the new Pathways system to orchestrate distributed computation for accelerators. In “PaLM: Scaling Language Modeling with Pathways”, we introduce the Pathways Language Model (PaLM), a 540-billion parameter, dense decoder-only Transformer model trained with the Pathways system, which enabled us to efficiently train a single model across multiple TPU v4 Pods. We evaluated PaLM on hundreds of language understanding and generation tasks, and found that it achieves state-of-the-art few-shot performance across most tasks, by significant margins in many cases.

Apr 5, 2022

Why Quantum Computing Is Closer Than You Think

Posted by in categories: computing, quantum physics

When I first started looking into quantum computing I had a fairly pessimistic view about its near-term commercial prospects, but I’ve come to think we’re only a few years away from seeing serious returns on the technology.

Apr 5, 2022

CRISPR-free base editors with enhanced activity and expanded targeting scope in mitochondrial and nuclear DNA

Posted by in categories: biotech/medical, evolution

Continuously evolved double-stranded DNA deaminases increase the efficiency and targeting scope of all-protein base editors.

Apr 5, 2022

Google, NYU & Maryland U’s Token-Dropping Approach Reduces BERT Pretraining Time by 25%

Posted by in category: robotics/AI

The pretraining of BERT-type large language models — which can scale up to billions of parameters — is crucial for obtaining state-of-the-art performance on many natural language processing (NLP) tasks. This pretraining process however is expensive, and has become a bottleneck hindering the industrial application of such large language models.

In the new paper Token Dropping for Efficient BERT Pretraining, a research team from Google, New York University, and the University of Maryland proposes a simple but effective “token dropping” technique that significantly reduces the pretraining cost of transformer models such as BERT, without degrading performance on downstream fine-tuning tasks.

The team summarizes their main contributions as:

Apr 5, 2022

A new approach that could improve how robots interact in conversational groups

Posted by in categories: biotech/medical, robotics/AI

To effectively interact with humans in crowded social settings, such as malls, hospitals, and other public spaces, robots should be able to actively participate in both group and one-to-one interactions. Most existing robots, however, have been found to perform much better when communicating with individual users than with groups of conversing humans.

Hooman Hedayati and Daniel Szafir, two researchers at University of North Carolina at Chapel Hill, have recently developed a new data-driven technique that could improve how robots communicate with groups of humans. This method, presented in a paper presented at the 2022 ACM/IEEE International Conference on Human-Robot Interaction (HRI ‘22), allows robots to predict the positions of humans in conversational groups, so that they do not mistakenly ignore a person when their sensors are fully or partly obstructed.

Continue reading “A new approach that could improve how robots interact in conversational groups” »

Apr 4, 2022

Scientists accidentally discover first-ever “animal” that doesn’t need oxygen to breathe

Posted by in category: futurism

O,.o! Circa 2020


The discovery expands the definition of what an “animal” can be, researchers say.

Apr 4, 2022

Secret of radiation-proof bugs proposed

Posted by in category: futurism

Circa 2004


Internal antioxidants may shield cells from radiation damage.

Apr 4, 2022

A million times faster: DNA nanotechnology could speed up pharmaceutical development while minimizing costs

Posted by in categories: biotech/medical, chemistry, economics, nanotechnology

A new tool speeds up development of vaccines and other pharmaceutical products by more than 1 million times while minimizing costs.

In search of pharmaceutical agents such as new vaccines, industry will routinely scan thousands of related candidate molecules. A novel technique allows this to take place on the nano scale, minimizing use of materials and energy. The work is published in the journal Nature Chemistry.

More than 40,000 molecules can be synthesized and analyzed within an area smaller than a pinhead. The method, developed through a highly interdisciplinary research effort in Denmark, promises to drastically reduce the amounts of material, energy, and economic cost for .

Apr 4, 2022

10 Difficult Problems Quantum Computers can Solve Easily

Posted by in categories: computing, quantum physics

ab initio calculations

Classical computing is of very little help when the task to be accomplished pertains to ab initio calculations. With quantum computing in place, you have a quantum system simulating another quantum system. Furthermore, tasks such as modelling atomic bonding or estimating electron orbital overlaps can be done much more precisely.

Apr 4, 2022

AMD Radeon Software Reportedly Alters CPU Settings Without User Knowledge

Posted by in category: computing

Adrenalin drivers seemingly go haywire.


German publication Igor’s Lab investigated a potential abnormality surrounding AMD’s Adrenalin GPU software. The current speculation is that the integration of AMD’s Ryzen Master module into the Adrenalin version 22.3.1 software or later has inadvertently allowed the program to manipulate CPU PBO and Precision Boost settings without the user’s permission.

According to user feedback, the problem occurs only with AMD CPU and GPU combinations since that hardware supports the Ryzen Master SDK. So if you are running a system that has either an Intel CPU paired to a Radeon GPU, or a Ryzen CPU paired to an Nvidia GPU, you will not encounter this problem.