Archive for the ‘Quantum Computer’ Category

The quantum race is here, and Colorado is poised to lead the pack – 9News.com KUSA

The quantum race is here, and Colorado is poised to lead the pack  9News.com KUSA

Go here to see the original:
The quantum race is here, and Colorado is poised to lead the pack - 9News.com KUSA

The Problem with Quantum Computers – Scientific American Blog Network

By now, most people have heard that quantum computing is a revolutionary technology that leverages the bizarre characteristics of quantum mechanics to solve certain problems faster than regular computers can. Those problems range from the worlds of mathematics to retail business, and physics to finance. If we get quantum technology right, the benefits should lift the entire economy and enhance U.S. competitiveness.

The promise of quantum computing was first recognized in the 1980s yet remains unfulfilled. Quantum computers are exceedingly difficult to engineer, buildand program. As a result, they are crippled by errors in the form of noise, faultsand loss of quantum coherence, which is crucial to their operation and yet falls apart before any nontrivial program has a chance to run to completion.

This loss of coherence (called decoherence), caused by vibrations, temperature fluctuations, electromagnetic waves and other interactions with the outside environment, ultimately destroys the exotic quantum properties of the computer. Given the current pervasiveness of decoherence and other errors, contemporary quantum computers are unlikely to return correct answers for programs of even modest execution time.

While competing technologies and competing architectures are attacking these problems, no existing hardware platform can maintain coherence and provide the robust error correction required for large-scale computation. A breakthrough is probably several years away.

The billion-dollar question in the meantime is, how do we get useful results out of a computer that becomes unusably unreliable before completing a typical computation?

Answers are coming from intense investigation across a number of fronts, with researchers in industry, academia and the national laboratories pursuing a variety of methods for reducing errors. One approach is to guess what an error-free computation would look like based on the results of computations with various noise levels. A completely different approach, hybrid quantum-classical algorithms, runs only the most performance-critical sections of a program on a quantum computer, with the bulk of the program running on a more robust classical computer. These strategies and others are proving to be useful for dealing with the noisy environment of todays quantum computers.

While classical computers are also affected by various sources of errors, these errors can be corrected with a modest amount of extra storage and logic. Quantum error correction schemes do exist but consume such a large number of qubits (quantum bits) that relatively few qubits remain for actual computation. That reduces the size of the computing task to a tiny fraction of what could run on defect-free hardware.

To put in perspective the importance of being stingy with qubit consumption, todays state-of-the-art, gate-based quantum computers, which use logic gates analogous to those forming the digital circuits found in the computer, smartphoneor tablet youre reading this article on, boast a mere 50 qubits. That is just a tiny fraction of the number of classical bits your device has available to it, typically hundreds of billions.

TAMING DEFECTS TO GET SOMETHING DONE

The trouble is, quantum mechanics challenges our intuition. So we struggle to figure out the best algorithms for performing meaningful tasks. To help overcome these problems, our team at Los Alamos National Laboratory is developing a method to invent and optimize algorithms that perform useful tasks on noisy quantum computers.

Algorithms are the lists of operations that tell a computer to do something, analogous to a cooking recipe. Compared to classical algorithms, the quantum kind are best kept as short as possible and, we have found, best tailored to the particular defects and noise regime of a given hardware device. That enables the algorithm to execute more processing steps within the constrained time frame before decoherence reduces the likelihood of a correct result to nearly zero.

In our interdisciplinary work on quantum computing at Los Alamos, funded by the Laboratory Directed Research and Development program, we are pursuing a key step in getting algorithms to run effectively. The main idea is to reduce the number of gates in an attempt to finish execution before decoherence and other sources of errors have a chance to unacceptably reduce the likelihood of success.

We use machine learning to translate, or compile, a quantum circuit into an optimally short equivalent that is specific to a particular quantum computer. Until recently, we have employed machine-learning methods on classical computers to search for shortened versions of quantum programs. Now, in a recent breakthrough, we have devised an approach that uses currently available quantum computers to compile their own quantum algorithms. That will avoid the massive computational overhead required to simulate quantum dynamics on classical computers.

Because this approach yields shorter algorithms than the state of the art, they consequently reduce the effects of noise. This machine-learning approach can also compensate for errors in a manner specific to the algorithm and hardware platform. It might find, for instance, that one qubit is less noisy than another, so the algorithm preferentially uses better qubits. In that situation, the machine learning creates a general algorithm to compute the assigned task on that computer using the fewest computational resources and the fewest logic gates. Thus optimized, the algorithm can run longer.

This method, which has worked in a limited setting on quantum computers now available to the public on the cloud, also takes advantage of quantum computers superior ability to scale-up algorithms for large problems on the larger quantum computers envisioned for the future.

New work with quantum algorithms will give both experts and nonexperts the tools to perform calculations on a quantum computer. Application developers can begin to take advantage of quantum computings potential for accelerating execution speed beyond the limits of conventional computing. These advances may bring us all several steps closer to having robust, reliable large-scale quantum computers to solve complex real-world problems that bring even the fastest classical computers to their knees.

Read more:
The Problem with Quantum Computers - Scientific American Blog Network

MIT researchers use quantum computing to observe entanglement – MIT News

  1. MIT researchers use quantum computing to observe entanglement  MIT News
  2. Did physicists create a wormhole in a quantum computer?  Nature.com
  3. A 'Wormhole' Built on a Quantum Computer Teleported Information as Predicted  ScienceAlert
  4. Google's quantum computer suggests that wormholes are real  Big Think
  5. A quantum computer has simulated a wormhole for the first time  New Scientist
  6. View Full Coverage on Google News

Follow this link:
MIT researchers use quantum computing to observe entanglement - MIT News

Post-Quantum Cryptography | CSRC – NIST

The Candidates to be Standardized and Round 4 Submissions were announced July 5, 2022. NISTIR 8413, Status Report on the Third Round of the NIST Post-Quantum Cryptography Standardization Process is now available.

NIST initiated a process to solicit, evaluate, and standardize one or more quantum-resistant public-key cryptographic algorithms.Full details can be found in the Post-Quantum Cryptography Standardization page.

In recent years, there has been a substantial amount of research on quantum computers machines that exploit quantum mechanical phenomena to solve mathematical problems that are difficult or intractable for conventional computers. If large-scale quantum computers are ever built, they will be able to break many of the public-key cryptosystems currently in use. This would seriously compromise the confidentiality and integrity of digital communications on the Internet and elsewhere. The goal of post-quantum cryptography (also called quantum-resistant cryptography) is to develop cryptographic systems that are secure against both quantum and classical computers, and can interoperate with existing communications protocols and networks.

The question of when a large-scale quantum computer will be built is a complicated one. While in the past it was less clear that large quantum computers are a physical possibility, many scientists now believe it to be merely a significant engineering challenge. Some engineers even predict that within the next twenty or so years sufficiently large quantum computers will be built to break essentially all public key schemes currently in use. Historically, it has taken almost two decades to deploy our modern public key cryptography infrastructure. Therefore, regardless of whether we can estimate the exact time of the arrival of the quantum computing era, we must begin now to prepare our information security systems to be able to resist quantum computing.

Here is the original post:
Post-Quantum Cryptography | CSRC - NIST

Quantum chemistry – Wikipedia

Chemistry based on quantum physics

Quantum chemistry, also called molecular quantum mechanics, is a branch of physical chemistry focused on the application of quantum mechanics to chemical systems, particularly towards the quantum-mechanical calculation of electronic contributions to physical and chemical properties of molecules, materials, and solutions at the atomic level. These calculations include systematically applied approximations intended to make calculations computationally feasible while still capturing as much information about important contributions to the computed wave functions as well as to observable properties such as structures, spectra, and thermodynamic properties. Quantum chemistry is also concerned with the computation of quantum effects on molecular dynamics and chemical kinetics.

Chemists rely heavily on spectroscopy through which information regarding the quantization of energy on a molecular scale can be obtained. Common methods are infra-red (IR) spectroscopy, nuclear magnetic resonance (NMR) spectroscopy, and scanning probe microscopy. Quantum chemistry may be applied to the prediction and verification of spectroscopic data as well as other experimental data.

Many quantum chemistry studies are focused on the electronic ground state and excited states of individual atoms and molecules as well as the study of reaction pathways and transition states that occur during chemical reactions. Spectroscopic properties may also be predicted. Typically, such studies assume the electronic wave function is adiabatically parameterized by the nuclear positions (i.e., the BornOppenheimer approximation). A wide variety of approaches are used, including semi-empirical methods, density functional theory, Hartree-Fock calculations, quantum Monte Carlo methods, and coupled cluster methods.

Understanding electronic structure and molecular dynamics through the development of computational solutions to the Schrdinger equation is a central goal of quantum chemistry. Progress in the field depends on overcoming several challenges, including the need to increase the accuracy of the results for small molecular systems, and to also increase the size of large molecules that can be realistically subjected to computation, which is limited by scaling considerations the computation time increases as a power of the number of atoms.

Some view the birth of quantum chemistry as starting with the discovery of the Schrdinger equation and its application to the hydrogen atom in 1926.[citation needed] However, the 1927 article of Walter Heitler (19041981) and Fritz London, is often recognized as the first milestone in the history of quantum chemistry. This is the first application of quantum mechanics to the diatomic hydrogen molecule, and thus to the phenomenon of the chemical bond. In the following years much progress was accomplished by Robert S. Mulliken, Max Born, J. Robert Oppenheimer, Linus Pauling, Erich Hckel, Douglas Hartree, Vladimir Fock, to cite a few. The history of quantum chemistry also goes through the 1838 discovery of cathode rays by Michael Faraday, the 1859 statement of the black-body radiation problem by Gustav Kirchhoff, the 1877 suggestion by Ludwig Boltzmann that the energy states of a physical system could be discrete, and the 1900 quantum hypothesis by Max Planck that any energy radiating atomic system can theoretically be divided into a number of discrete energy elements such that each of these energy elements is proportional to the frequency with which they each individually radiate energy and a numerical value called Planck's constant. Then, in 1905, to explain the photoelectric effect (1839), i.e., that shining light on certain materials can function to eject electrons from the material, Albert Einstein postulated, based on Planck's quantum hypothesis, that light itself consists of individual quantum particles, which later came to be called photons (1926). In the years to follow, this theoretical basis slowly began to be applied to chemical structure, reactivity, and bonding. Probably the greatest contribution to the field was made by Linus Pauling.[citation needed]

The first step in solving a quantum chemical problem is usually solving the Schrdinger equation (or Dirac equation in relativistic quantum chemistry) with the electronic molecular Hamiltonian. This is called determining the electronic structure of the molecule. It can be said that the electronic structure of a molecule or crystal implies essentially its chemical properties. An exact solution for the Schrdinger equation can only be obtained for the hydrogen atom (though exact solutions for the bound state energies of the hydrogen molecular ion have been identified in terms of the generalized Lambert W function). Since all other atomic, or molecular systems, involve the motions of three or more "particles", their Schrdinger equations cannot be solved exactly and so approximate solutions must be sought.

Although the mathematical basis of quantum chemistry had been laid by Schrdinger in 1926, it is generally accepted that the first true calculation in quantum chemistry was that of the German physicists Walter Heitler and Fritz London on the hydrogen (H2) molecule in 1927.[citation needed] Heitler and London's method was extended by the American theoretical physicist John C. Slater and the American theoretical chemist Linus Pauling to become the valence-bond (VB) [or HeitlerLondonSlaterPauling (HLSP)] method. In this method, attention is primarily devoted to the pairwise interactions between atoms, and this method therefore correlates closely with classical chemists' drawings of bonds. It focuses on how the atomic orbitals of an atom combine to give individual chemical bonds when a molecule is formed, incorporating the two key concepts of orbital hybridization and resonance.

An alternative approach was developed in 1929 by Friedrich Hund and Robert S. Mulliken, in which electrons are described by mathematical functions delocalized over an entire molecule. The HundMulliken approach or molecular orbital (MO) method is less intuitive to chemists, but has turned out capable of predicting spectroscopic properties better than the VB method. This approach is the conceptual basis of the HartreeFock method and further post HartreeFock methods.

The ThomasFermi model was developed independently by Thomas and Fermi in 1927. This was the first attempt to describe many-electron systems on the basis of electronic density instead of wave functions, although it was not very successful in the treatment of entire molecules. The method did provide the basis for what is now known as density functional theory (DFT). Modern day DFT uses the KohnSham method, where the density functional is split into four terms; the KohnSham kinetic energy, an external potential, exchange and correlation energies. A large part of the focus on developing DFT is on improving the exchange and correlation terms. Though this method is less developed than post HartreeFock methods, its significantly lower computational requirements (scaling typically no worse than n3 with respect to n basis functions, for the pure functionals) allow it to tackle larger polyatomic molecules and even macromolecules. This computational affordability and often comparable accuracy to MP2 and CCSD(T) (post-HartreeFock methods) has made it one of the most popular methods in computational chemistry.

A further step can consist of solving the Schrdinger equation with the total molecular Hamiltonian in order to study the motion of molecules. Direct solution of the Schrdinger equation is called quantum dynamics, whereas its solution within the semiclassical approximation is called semiclassical dynamics. Purely classical simulations of molecular motion are referred to as molecular dynamics (MD). Another approach to dynamics is a hybrid framework known as mixed quantum-classical dynamics; yet another hybrid framework uses the Feynman path integral formulation to add quantum corrections to molecular dynamics, which is called path integral molecular dynamics. Statistical approaches, using for example classical and quantum Monte Carlo methods, are also possible and are particularly useful for describing equilibrium distributions of states.

In adiabatic dynamics, interatomic interactions are represented by single scalar potentials called potential energy surfaces. This is the BornOppenheimer approximation introduced by Born and Oppenheimer in 1927. Pioneering applications of this in chemistry were performed by Rice and Ramsperger in 1927 and Kassel in 1928, and generalized into the RRKM theory in 1952 by Marcus who took the transition state theory developed by Eyring in 1935 into account. These methods enable simple estimates of unimolecular reaction rates from a few characteristics of the potential surface.

Non-adiabatic dynamics consists of taking the interaction between several coupled potential energy surface (corresponding to different electronic quantum states of the molecule). The coupling terms are called vibronic couplings. The pioneering work in this field was done by Stueckelberg, Landau, and Zener in the 1930s, in their work on what is now known as the LandauZener transition. Their formula allows the transition probability between two diabatic potential curves in the neighborhood of an avoided crossing to be calculated. Spin-forbidden reactions are one type of non-adiabatic reactions where at least one change in spin state occurs when progressing from reactant to product.

Read more:
Quantum chemistry - Wikipedia