New! Sign up for our free email newsletter.
Science News
from research organizations

The right squeeze for quantum computing

Date:
May 31, 2018
Source:
Hokkaido University
Summary:
A new theoretical model involving squeezing light to just the right amount to accurately transmit information using subatomic particles is bringing us closer to a new era of computing.
Share:
FULL STORY

A new theoretical model involving squeezing light to just the right amount to accurately transmit information using subatomic particles is bringing us closer to a new era of computing.

Scientists at Hokkaido University and Kyoto University have developed a theoretical approach to quantum computing that is 10 billion times more tolerant to errors than current theoretical models. Their method brings us closer to developing quantum computers that use the diverse properties of subatomic particles to transmit, process and store extremely large amounts of complex information.

Quantum computing has the potential to solve problems involving vast amounts of information, such as modelling complex chemical processes, far better and faster than modern computers.

Computers currently store data by coding it into "bits." A bit can exist in one of two states: 0 and 1. Scientists have been investigating ways to employ subatomic particles, called "quantum bits," which can exist in more than just two separate states, for the storage and processing of much vaster amounts of information. Quantum bits are the building blocks of quantum computers.

One such approach involves using the inherent properties in photons of light, such as encoding information as quantum bits into a light beam by digitizing patterns of the electromagnetic field. But the encoded information can be lost from light waves during quantum computation, leading to an accumulation of errors. To reduce information loss, scientists have been experimenting with "squeezing" light. Squeezing is a process that removes tiny quantum-level fluctuations, referred to as noise, from an electromagnetic field. Noise introduces a certain level of uncertainty into the amplitude and phase of the electromagnetic field. Squeezing is thus an efficient tool for the optical implementation of quantum computers, but the current usage is inadequate.

In a paper published in the journal Physical Review X, Akihisa Tomita, an applied physicist at Hokkaido University, and his colleagues suggested a novel way to dramatically reduce errors when using this approach. They developed a theoretical model that uses both the properties of quantum bits and the modes of the electromagnetic field in which they exist. The approach involves squeezing light by removing error-prone quantum bits, when quantum bits cluster together.

This model is ten billion times more tolerant to errors than current experimental methods, meaning that it tolerates up to one error every 10,000 calculations.

"The approach is achievable using currently available technologies, and could further advance developments in quantum computing research," says Akihisa Tomita of Hokkaido University.


Story Source:

Materials provided by Hokkaido University. Note: Content may be edited for style and length.


Journal Reference:

  1. Kosuke Fukui, Akihisa Tomita, Atsushi Okamoto, Keisuke Fujii. High-Threshold Fault-Tolerant Quantum Computation with Analog Quantum Error Correction. Physical Review X, 2018; 8 (2) DOI: 10.1103/PhysRevX.8.021054

Cite This Page:

Hokkaido University. "The right squeeze for quantum computing." ScienceDaily. ScienceDaily, 31 May 2018. <www.sciencedaily.com/releases/2018/05/180531102801.htm>.
Hokkaido University. (2018, May 31). The right squeeze for quantum computing. ScienceDaily. Retrieved December 21, 2024 from www.sciencedaily.com/releases/2018/05/180531102801.htm
Hokkaido University. "The right squeeze for quantum computing." ScienceDaily. www.sciencedaily.com/releases/2018/05/180531102801.htm (accessed December 21, 2024).

Explore More

from ScienceDaily

RELATED STORIES