Toggle light / dark theme

Quicker time to discovery. That’s what scientists focused on quantum chemistry are looking for. According to Bert de Jong, Computational Chemistry, Materials and Climate Group Lead, Computational Research Division, Lawrence Berkeley National Lab (LBNL), “I’m a computational chemist working extensively with experimentalists doing interdisciplinary research. To shorten time to scientific discovery, I need to be able to run simulations at near-real-time, or at least overnight, to drive or guide the next experiments.” Changes must be made in the HPC software used in quantum chemistry research to take advantage of advanced HPC systems to meet the research needs of scientists both today and in the future.

NWChem is a widely used open source software computational chemistry package that includes both quantum chemical and molecular dynamics functionality. The NWChem project started around the mid-1990s, and the code was designed from the beginning to take advantage of parallel computer systems. NWChem is actively developed by a consortium of developers and maintained by the Environmental Molecular Sciences Laboratory (EMSL) located at the Pacific Northwest National Laboratory (PNNL) in Washington State. NWChem aims to provide its users with computational chemistry tools that are scalable both in their ability to treat large scientific computational chemistry problems efficiently, and in their use of available parallel computing resources from high-performance parallel supercomputers to conventional workstation clusters.

“Rapid evolution of the computational hardware also requires significant effort geared toward the modernization of the code to meet current research needs,” states Karol Kowalski, Capability Lead for NWChem Development at PNNL.

Read more

A few years ago, researchers from Germany and Japan were able to simulate one percent of human brain activity for a single second. It took the processing power of one of the world’s most powerful supercomputers to make that happen.

Hands down, the human brain is by far the most powerful, energy efficient computer ever created.

So what if we could harness the power of the human brain by using actual brain cells to power the next generation of computers?

As crazy as it sounds, that’s exactly what neuroscientist Osh Agabi is building. Koniku, Agabi’s startup, has developed a prototype 64-neuron silicon chip.

Read more

Solving the turbulence plasma mystery.


Cutting-edge simulations run at Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center (NERSC) over a two-year period are helping physicists better understand what influences the behavior of the plasma turbulence that is driven by the intense heating necessary to create fusion energy. This research has yielded exciting answers to long-standing questions about plasma heat loss that have previously stymied efforts to predict the performance of fusion reactors and could help pave the way for this alternative energy source.

The key to making fusion work is to maintain a sufficiently high temperature and density to enable the atoms in the reactor to overcome their mutual repulsion and bind to form helium. But one side effect of this process is turbulence, which can increase the rate of plasma, significantly limiting the resulting energy output. So researchers have been working to pinpoint both what causes the turbulence and how to control or possibly eliminate it.

Because are extremely complex and expensive to design and build, supercomputers have been used for more than 40 years to simulate the conditions to create better reactor designs. NERSC is a Department of Energy Office of Science User Facility that has supported fusion research since 1974.

In what appears at first to be a storyline ripped from a sci-fi thriller, a multi-national research team spread across two continents, four countries, and ten years in the making have created a model of a supercomputer that runs on the same substance that living things use as an energy source.

Humans and virtually all living things rely on Adenosine triphosphate ( ATP ) to provide the energy our cells need to perform daily functions. The biological computer created by the team led by Professor Dan Nicolau, Chair of the Department of Bioengineering at McGill, also relies on ATP for power.

The biological computer is able to process information very quickly and operates accurately using parallel networks like contemporary massive electronic super computers. In addition, the model is lot smaller in size, uses relatively less energy, and functions using proteins that are present in all living cells.

Read more

I must admit, when people see that you work with Quantum Computing and/ or networking; they have no idea how to classify you because you’re working on Nextgen “disruptive” technology that most of mainstream has not been exposed to.


Peter Wittek and I met more than a decade ago while he was an exchange student in Singapore. I consider him one of the most interesting people I’ve met and an inspiration to us all.

Currently, he is a research scientist working on quantum machine learning, an emergent field halfway between data science and quantum information processing. Peter also has a long history in machine learning on supercomputers and large-scale simulations of quantum systems. As a former digital nomad, Peter has been to over a hundred countries, he is currently based in Barcelona where, outside work hours, he focuses on dancing salsa, running long distances, and advising startups.

2016-02-28-1456691593-1100960-IMG_2684.jpg

Read more

Researchers have taken on the problem of reducing a super computer the size of a basketball field to that of a book. The answer is “biocomputers” – incredibly powerful machines capable of performing multiple calculations with a fraction of energy.

According to study coordinator Heiner Linke, who heads nanoscience at Lund University in Sweden, “a biocomputer requires less than one percent of the energy an electronic transistor needs to carry out one calculation step.”

A biocomputer is useful because ordinary computers are incapable of solving combinational problems, such as those dealing with cryptography or other tasks requiring that a multitude of possible solutions be considered before deciding on the optimal one. These already exist, but the new research from Lund tackles the key problems of scalability and energy efficiency.

Read more