Toggle light / dark theme

Meeting the world’s energy demands is reaching a critical point. Powering the technological age has caused issues globally. It is increasingly important to create superconductors that can operate at ambient pressure and temperature. This would go a long way toward solving the energy crisis.

Advancements with superconductivity hinge on advances in . When electrons inside of quantum materials undergo a phase transition, the electrons can form intricate patterns, such as fractals. A fractal is a never-ending pattern. When zooming in on a fractal, the image looks the same. Commonly seen fractals can be a tree or frost on a windowpane in winter. Fractals can form in two dimensions, like the frost on a window, or in three-dimensional space like the limbs of a tree.

Dr. Erica Carlson, a 150th Anniversary Professor of Physics and Astronomy at Purdue University, led a team that developed theoretical techniques for characterizing the fractal shapes that these electrons make, in order to uncover the underlying physics driving the patterns.

In a development that could make quantum computers less prone to errors, a team of physicists from Quantinuum, California Institute of Technology and Harvard University has created a signature of non-Abelian anyons (nonabelions) in a special type of quantum computer. The team has published their results on the arXiv preprint server.

As scientists work to design and build a truly useful quantum computer, one of the difficulties is trying to account for errors that creep in. In this new effort, the researchers have looked to anyons for help.

Anyons are quasiparticles that exist in two dimensions. They are not true particles, but instead exist as vibrations that act like particles—certain groups of them are called nonabelions. Prior research has found that nonabelions have a unique and useful property—they remember some of their own history. This property makes them potentially useful for creating less error-prone quantum computers. But creating, manipulating and doing useful things with them in a quantum computer is challenging. In this new work, the team have come close by creating a physical simulation of nonabelions in action.

O.o!!!


The unification of general relativity and quantum theory is one of the fascinating problems of modern physics. One leading solution is Loop Quantum Gravity (LQG). Simulating LQG may be important for providing predictions which can then be tested experimentally. However, such complex quantum simulations cannot run efficiently on classical computers, and quantum computers or simulators are needed. Here, we experimentally demonstrate quantum simulations of spinfoam amplitudes of LQG on an integrated photonics quantum processor. We simulate a basic transition of LQG and show that the derived spinfoam vertex amplitude falls within 4% error with respect to the theoretical prediction, despite experimental imperfections.

During its ongoing Think 2023 conference, IBM today announced an end-to-end solution to prepare organisations to adopt quantum-safe cryptography. Called Quantum Safe technology, it is a set of tools and capabilities that integrates IBM’s deep security expertise. Quantum-safe cryptography is a technique to identify algorithms that are resistant to attacks by both classical and quantum computers.

Under Quantum Safe technology, IBM is offering three capabilities. First is the Quantum Safe Explorer to locate cryptographic assets, dependencies, and vulnerabilities and aggregate all potential risks in one central location. Next is the Quantum Safe Advisor which allows the creation of a cryptographic inventory to prioritise risks. Lastly, the Quantum Safe Remidiator lets organisations test quantum-safe remediation patterns and deploy quantum-safe solutions.

In addition, the company has also announced IBM Safe Roadmap, which will serve as the guide for industries to adopt quantum technology. IBM Quantum Safe Roadmap is the company’s first blueprint to help companies in dealing with anticipated cryptographic standards and requirements and protect systems from vulnerabilities.

For two decades, physicists have tried to directly manipulate the spin of electrons in 2D materials like graphene. Doing so could spark key advances in the burgeoning world of 2D electronics, a field where super-fast, small and flexible electronic devices carry out computations based on quantum mechanics.

Standing in the way is that the typical way in which scientists measure the spin of electrons—an essential behavior that gives everything in the physical universe its structure—usually doesn’t work in 2D materials. This makes it incredibly difficult to fully understand the materials and propel forward technological advances based on them. But a team of scientists led by Brown University researchers believe they now have a way around this longstanding challenge. They describe their solution in a new study published in Nature Physics.

In the study, the team—which also include scientists from the Center for Integrated Nanotechnologies at Sandia National Laboratories, and the University of Innsbruck—describe what they believe to be the first measurement showing direct interaction between electrons spinning in a 2D material and photons coming from microwave radiation.

A new experiment uses superconducting qubits to demonstrate that quantum mechanics violates what’s called local realism by allowing two objects to behave as a single quantum system no matter how large the separation between them. The experiment wasn’t the first to show that local realism isn’t how the Universe works—it’s not even the first to do so with qubits.

But it’s the first to separate the qubits by enough distance to ensure that light isn’t fast enough to travel between them while measurements are made. And it did so by cooling a 30-meter-long aluminum wire to just a few milliKelvin. Because the qubits are so easy to control, the experiment provides a new precision to these sorts of measurements. And the hardware setup may be essential for future quantum computing efforts.

The silicon microchips of future quantum computers will be packed with millions, if not billions of qubits—the basic units of quantum information—to solve the greatest problems facing humanity. And with millions of qubits needing millions of wires in the microchip circuitry, it was always going to get cramped in there.

But now engineers at UNSW Sydney have made an important step toward solving a long-standing problem about giving their more breathing space—and it all revolves around jellybeans.

Not the kind we rely on for a sugar hit to get us past the 3pm slump. But jellybean quantum dots—elongated areas between qubit pairs that create more space for wiring without interrupting the way the paired qubits interact with each other.

In the United States, the first step on the road to exascale HPC systems began with a series of workshops in 2007. It wasn’t until a decade and a half later that the 1,686 petaflops “Frontier” system at Oak Ridge National Laboratory went online. This year, Argonne National Laboratory is preparing for the switch to be turned on for “Aurora,” which will be either the second or the third such exascale machine in the United States, depending on the timing of the “El Capitan” system at Lawrence Livermore National Laboratory.

There were delays and setbacks on the road to exascale for all of these machines, as well as technology changes, ongoing competition with China, and other challenges. But don’t expect the next leap to zettascale – or even quantum computing – to be any quicker, according to Rick Stevens, associate laboratory director of computing for environment and life sciences at Argonne. Both could take another 15 to 20 years or more.

Such is the nature of HPC.

Year 2020 face_with_colon_three


A trio of theoretical physicists at the Pennsylvania State University has calculated the upper limit for the possible quantization of time—they suggest 10−33 seconds as the upper limit for the period of a universal oscillator. In their paper published in the journal Physical Review Letters, Garrett Wendel, Luis Martínez and Martin Bojowald outline their theory and suggest a possible way to prove it.

For many years, have been trying to explain a major problem—the suggests that time is a continuous quantity, one that can move slower or faster depending on acceleration and gravity conditions. But quantum mechanics theories suggest that time ticks away at a steady pace, like the frames of a movie being played out. In this scenario, time must be universal. For both theories to be right, this contradiction must be explained in a rational way.

Some theorists have suggested that one possible explanation for the apparent discrepancy is that time can be quantized as spacetime, similar to theories describing quantum gravity. In such a scenario, spacetime is not described as continuous, but is instead divided into smaller units, which would by necessity have to correspond to the Planck length. This is, of course, far too small to be detectable. The would also require that such discrete packets of time would each expire. This scenario suggests there would need to be a universal clock that ticks away at a very small unit of time. And under this scenario, universal time would exist throughout the universe and also interact with matter. It also raises the question of how fast would such a clock tick.