Archive for the ‘Quantum Computing’ Category

Explainer: What is a quantum computer? | MIT Technology Review

They wont wipe out conventional computers, though. Using a classical machine will still be the easiest and most economical solution for tackling most problems. But quantum computers promise to power exciting advances in various fields, from materials science to pharmaceuticals research. Companies are already experimenting with them to develop things like lighter and more powerful batteries for electric cars, and to help create novel drugs.

The secret to a quantum computers power lies in its ability to generate and manipulate quantum bits, or qubits.

Today's computers use bitsa stream of electrical or optical pulses representing1s or0s. Everything from your tweets and e-mails to your iTunes songs and YouTube videos are essentially long strings of these binary digits.

Quantum computers, on the other hand, usequbits, whichare typically subatomic particles such as electrons or photons. Generating and managing qubits is a scientific and engineering challenge. Some companies, such as IBM, Google, and Rigetti Computing, use superconducting circuits cooled to temperatures colder than deep space. Others, like IonQ, trap individual atoms in electromagnetic fields on a silicon chip in ultra-high-vacuum chambers. In both cases, the goal is to isolate the qubits in a controlled quantum state.

Qubits have some quirky quantum properties that mean a connected group of them can provide way more processing power than the same number of binary bits. One of those properties is known as superposition and another is called entanglement.

Qubits can represent numerous possible combinations of 1and 0 at the same time. This ability to simultaneously be in multiple states is called superposition. To put qubits into superposition, researchers manipulate them using precision lasers or microwave beams.

Read the rest here:
Explainer: What is a quantum computer? | MIT Technology Review

What Is Quantum Computing? | NVIDIA Blog

Twenty-seven years before Steve Jobs unveiled a computer you could put in your pocket, physicist Paul Benioff published a paper showing it was theoretically possible to build a much more powerful system you could hide in a thimble a quantum computer.

Named for the subatomic physics it aimed to harness, the concept Benioff described in 1980 still fuels research today, including efforts to build the next big thing in computing: a system that could make a PC look in some ways quaint as an abacus.

Richard Feynman a Nobel Prize winner whose wit-laced lectures brought physics to a broad audience helped establish the field, sketching out how such systems could simulate quirky quantum phenomena more efficiently than traditional computers. So,

Quantum computing is a sophisticated approach to making parallel calculations, using the physics that governs subatomic particles to replace the more simplistic transistors in todays computers.

Quantum computers calculate using qubits, computing units that can be on, off or any value between, instead of the bits in traditional computers that are either on or off, one or zero. The qubits ability to live in the in-between state called superposition adds a powerful capability to the computing equation, making quantum computers superior for some kinds of math.

Using qubits, quantum computers could buzz through calculations that would take classical computers a loooong time if they could even finish them.

For example, todays computers use eight bits to represent any number between 0 and 255. Thanks to features like superposition, a quantum computer can use eight qubits to represent every number between 0 and 255, simultaneously.

Its a feature like parallelism in computing: All possibilities are computed at once rather than sequentially, providing tremendous speedups.

So, while a classical computer steps through long division calculations one at a time to factor a humongous number, a quantum computer can get the answer in a single step. Boom!

That means quantum computers could reshape whole fields, like cryptography, that are based on factoring what are today impossibly large numbers.

That could be just the start. Some experts believe quantum computers will bust through limits that now hinder simulations in chemistry, materials science and anything involving worlds built on the nano-sized bricks of quantum mechanics.

Quantum computers could even extend the life of semiconductors by helping engineers create more refined simulations of the quantum effects theyre starting to find in todays smallest transistors.

Indeed, experts say quantum computers ultimately wont replace classical computers, theyll complement them. And some predict quantum computers will be used as accelerators much as GPUs accelerate todays computers.

Dont expect to build your own quantum computer like a DIY PC with parts scavenged from discount bins at the local electronics shop.

The handful of systems operating today typically require refrigeration that creates working environments just north of absolute zero. They need that computing arctic to handle the fragile quantum states that power these systems.

In a sign of how hard constructing a quantum computer can be, one prototype suspends an atom between two lasers to create a qubit. Try that in your home workshop!

Quantum computing takes nano-Herculean muscles to create something called entanglement. Thats when two or more qubits exist in a single quantum state, a condition sometimes measured by electromagnetic waves just a millimeter wide.

Crank up that wave with a hair too much energy and you lose entanglement or superposition, or both. The result is a noisy state called decoherence, the equivalent in quantum computing of the blue screen of death.

A handful of companies such as Alibaba, Google, Honeywell, IBM, IonQ and Xanadu operate early versions of quantum computers today.

Today they provide tens of qubits. But qubits can be noisy, making them sometimes unreliable. To tackle real-world problems reliably, systems need tens or hundreds of thousands of qubits.

Experts believe it could be a couple decades before we get to a high-fidelity era when quantum computers are truly useful.

Predictions of when we reach so-called quantum computing supremacy the time when quantum computers execute tasks classical ones cant is a matter of lively debate in the industry.

The good news is the world of AI and machine learning put a spotlight on accelerators like GPUs, which can perform many of the types of operations quantum computers would calculate with qubits.

So, classical computers are already finding ways to host quantum simulations with GPUs today. For example, NVIDIA ran a leading-edge quantum simulation on Selene, our in-house AI supercomputer.

NVIDIA announced in the GTC keynote the cuQuantum SDK to speed quantum circuit simulations running on GPUs. Early work suggests cuQuantum will be able to deliver orders of magnitude speedups.

The SDK takes an agnostic approach, providing a choice of tools users can pick to best fit their approach. For example, the state vector method provides high-fidelity results, but its memory requirements grow exponentially with the number of qubits.

That creates a practical limit of roughly 50 qubits on todays largest classical supercomputers. Nevertheless weve seen great results (below) using cuQuantum to accelerate quantum circuit simulations that use this method.

Researchers from the Jlich Supercomputing Centre will provide a deep dive on their work with the state vector method in session E31941 at GTC (free with registration).

A newer approach, tensor network simulations, use less memory and more computation to perform similar work.

Using this method, NVIDIA and Caltech accelerated a state-of-the-art quantum circuit simulator with cuQuantum running on NVIDIA A100 Tensor Core GPUs. It generated a sample from a full-circuit simulation of the Google Sycamore circuit in 9.3 minutes on Selene, a task that 18 months ago experts thought would take days using millions of CPU cores.

Using the Cotengra/Quimb packages, NVIDIAs newly announced cuQuantum SDK, and the Selene supercomputer, weve generated a sample of the Sycamore quantum circuit at depth m=20 in record time less than 10 minutes, said Johnnie Gray, a research scientist at Caltech.

This sets the benchmark for quantum circuit simulation performance and will help advance the field of quantum computing by improving our ability to verify the behavior of quantum circuits, said Garnet Chan, a chemistry professor at Caltech whose lab hosted the work.

NVIDIA expects the performance gains and ease of use of cuQuantum will make it a foundational element in every quantum computing framework and simulator at the cutting edge of this research.

Sign up to show early interest in cuQuantum.

Learn more about quantum computing on the NVIDIA Technical Blog.

Visit link:
What Is Quantum Computing? | NVIDIA Blog

What’s next for quantum computing | MIT Technology Review

For years, quantum computings news cycle was dominated by headlines about record-setting systems. Researchers at Google and IBM have had spats over who achieved whatand whether it was worth the effort. But the time for arguing over whos got the biggest processor seems to have passed: firms are heads-down and preparing for life in the real world. Suddenly, everyone is behaving like grown-ups.

As if to emphasize how much researchers want to get off the hype train, IBM is expected to announce a processor in 2023 that bucks the trend of putting ever more quantum bits, or qubits, into play. Qubits, the processing units of quantum computers, can be built from a variety of technologies, including superconducting circuitry, trapped ions, and photons, the quantum particles of light.

IBM has long pursued superconducting qubits, and over the years the company has been making steady progress in increasing the number it can pack on a chip. In 2021, for example, IBM unveiled one with a record-breaking 127 of them. In November, it debuted its 433-qubit Osprey processor, and the company aims to release a 1,121-qubit processor called Condor in 2023.

But this year IBM is also expected to debut its Heron processor, which will have just 133 qubits. It might look like a backwards step, but as the company is keen to point out, Herons qubits will be of the highest quality. And, crucially, each chip will be able to connect directly to other Heron processors, heralding a shift from single quantum computing chips toward modular quantum computers built from multiple processors connected togethera move that is expected to help quantum computers scale up significantly.

Heron is a signal of larger shifts in the quantum computing industry. Thanks to some recent breakthroughs, aggressive roadmapping, and high levels of funding, we may see general-purpose quantum computers earlier than many would have anticipated just a few years ago, some experts suggest. Overall, things are certainly progressing at a rapid pace, says Michele Mosca, deputy director of the Institute for Quantum Computing at the University of Waterloo.

Here are a few areas where experts expect to see progress.

IBMs Heron project is just a first step into the world of modular quantum computing. The chips will be connected with conventional electronics, so they will not be able to maintain the quantumness of information as it moves from processor to processor. But the hope is that such chips, ultimately linked together with quantum-friendly fiber-optic or microwave connections, will open the path toward distributed, large-scale quantum computers with as many as a million connected qubits. That may be how many are needed to run useful, error-corrected quantum algorithms. We need technologies that scale both in size and in cost, so modularity is key, says Jerry Chow, director at IBMQuantum Hardware System Development.

More here:
What's next for quantum computing | MIT Technology Review

Could the Tech Industry Layoffs Be a Silver Lining for the Quantum Industry Cloud? – Quantum Computing Report

Could the Tech Industry Layoffs Be a Silver Lining for the Quantum Industry Cloud?  Quantum Computing Report

More:
Could the Tech Industry Layoffs Be a Silver Lining for the Quantum Industry Cloud? - Quantum Computing Report

Why IBM is no longer interested in breaking patent recordsand how it plans to measure innovation in the age of open source and quantum computing -…

Why IBM is no longer interested in breaking patent recordsand how it plans to measure innovation in the age of open source and quantum computing  Fortune

Continue reading here:
Why IBM is no longer interested in breaking patent recordsand how it plans to measure innovation in the age of open source and quantum computing -...