A selection of the most important recent news, articles, and papers about quantum computing.
News, Articles, and Analyses
(Tuesday, July 16, 2024) “We’re thrilled to announce the installation of our state-of-the-art neutral atom quantum computer at the National Quantum Computing Centre (NQCC). As the first company to deploy hardware under the NQCC’s quantum computing testbed programme, this milestone showcases our cutting-edge technology and de”
(Tuesday, July 16, 2024) “Oxford Ionics claim to have created the first quantum chip of its kind that could be mass-produced.”
(Wednesday, July 17, 2024) “CHICAGO (WCIA) — Illinois’ proposal to create a new quantum computing campus has a new partner with a federal agency. Governor J.B. Pritzker announced the partnership of Defense Advanced Research Projects Agency, part of the U.S. Department of Defense, with Illinois’ quantum computing campus Tuesday. The partnership is named Quantum Proving Ground. “The future of […]”
Technical Papers and Preprints
Authors: Buck, Samantha; Coleman, Robin; Sargsyan, Hayk
(Monday, July 05, 2021) “Quantum computing is usually associated with discrete quantum states and physical quantities possessing discrete eigenvalue spectrum. However, quantum computing in general is any computation accomplished by the exploitation of quantum properties of physical quantities, discrete or otherwise. It has been shown that physical quantities with continuous eigenvalue spectrum can be used for quantum computing as well. Currently, continuous variable quantum computing is a rapidly developing field both theoretically and experimentally. In this pedagogical introduction we present the basic theoretical concepts behind it and the tools for algorithm development. The paper targets readers with discrete quantum computing background, who are new to continuous variable quantum computing.”
(Thursday, May 25, 2023) “To fulfil the potential of quantum machine learning for practical applications in the near future, it needs to be robust against adversarial attacks. West and colleagues give an overview of recent developments in quantum adversarial machine learning, and outline key challenges and future research directions to advance the field.”
Authors: Kim, Youngseok; Govia, Luke C. G.; Dane, Andrew; Berg, Ewout van den; Zajac, David M.; Mitchell, Bradley; Liu, Yinyu; Balakrishnan, Karthik; Keefe, George; Stabile, Adam; Pritchett, Emily; Stehlik, Jiri; Kandala, Abhinav
(Tuesday, July 02, 2024) “Pre-fault tolerant quantum computers have already demonstrated the ability to estimate observable values accurately, at a scale beyond brute-force classical computation. This has been enabled by error mitigation techniques that often rely on a representative model on the device noise. However, learning and maintaining these models is complicated by fluctuations in the noise over unpredictable time scales, for instance, arising from resonant interactions between superconducting qubits and defect two-level systems (TLS). Such interactions affect the stability and uniformity of device performance as a whole, but also affect the noise model accuracy, leading to incorrect observable estimation. Here, we experimentally demonstrate that tuning of the qubit-TLS interactions helps reduce noise instabilities and consequently enables more reliable error-mitigation performance. These experiments provide a controlled platform for studying the performance of error mitigation in the presence of quasi-static noise. We anticipate that the capabilities introduced here will be crucial for the exploration of quantum applications on solid-state processors at non-trivial scales.”
Authors: Arnault, Pablo; Arrighi, Pablo; Herbert, Steven; Kasnetsi, Evi; Li, Tianyi
(Saturday, July 06, 2024) “We draw the current landscape of quantum algorithms, by classifying about 130 quantum algorithms, according to the fundamental mathematical problems they solve, their real-world applications, the main subroutines they employ, and several other relevant criteria. The primary objectives include revealing trends of algorithms, identifying promising fields for implementations in the NISQ era, and identifying the key algorithmic primitives that power quantum advantage.”
Share this:
- Threads
-
Categories Quantum Computing Tags algorithm, error mitigation, Infleqtion, machine learning, mathematics, neutral atom, NISQ, Oxford Ionics, quantum computing, qubit, superconducting
A selection of the most important recent news, articles, and papers about quantum computing.
News and Articles
(Wednesday, July 10, 2024) “Researchers at the University of Cologne have achieved a significant breakthrough in quantum materials, potentially setting the stage for advancements in topological superconductivity and robust quantum computing / publication in ‘Nature Physics’”
(Thursday, July 11, 2024) “UK industry and researchers will gain unparalleled access to the world’s most powerful quantum computers.”
(Friday, July 12, 2024) “Bob Sutor; Vice President and Practice Lead, Emerging Technologies, The Futurum Group will speak at IQT Quantum + AI in New York City October 29-30. Dr. Bob Sutor has been a technical leader and executive in the IT industry for over 40 years. He is a theoretical mathematician by training, with a Ph.D. from Princeton”
Technical Papers and Preprints
(Tuesday, June 25, 2024) “Fast, reliable logical operations are essential for the realization of useful quantum computers, as they are required to implement practical quantum algorithms at large scale. By redundantly encoding logical qubits into many physical qubits and using syndrome measurements to detect and subsequently correct errors, one can achieve very low logical error rates. However, for most practical quantum error correcting (QEC) codes such as the surface code, it is generally believed that due to syndrome extraction errors, multiple extraction rounds — on the order of the code distance d — are required for fault-tolerant computation. Here, we show that contrary to this common belief, fault-tolerant logical operations can be performed with constant time overhead for a broad class of QEC codes, including the surface code with magic state inputs and feed-forward operations, to achieve “algorithmic fault tolerance”. Through the combination of transversal operations and novel strategies for correlated decoding, despite only having access to partial syndrome information, we prove that the deviation from the ideal measurement result distribution can be made exponentially small in the code distance. We supplement this proof with circuit-level simulations in a range of relevant settings, demonstrating the fault tolerance and competitive performance of our approach. Our work sheds new light on the theory of fault tolerance, potentially reducing the space-time cost of practical fault-tolerant quantum computation by orders of magnitude.”
(Tuesday, July 02, 2024) “Quantum machine learning has gained considerable attention as quantum technology advances, presenting a promising approach for efficiently learning complex data patterns. Despite this promise, most contemporary quantum methods require significant resources for variational parameter optimization and face issues with vanishing gradients, leading to experiments that are either limited in scale or lack potential for quantum advantage. To address this, we develop a general-purpose, gradient-free, and scalable quantum reservoir learning algorithm that harnesses the quantum dynamics of neutral-atom analog quantum computers to process data. We experimentally implement the algorithm, achieving competitive performance across various categories of machine learning tasks, including binary and multi-class classification, as well as timeseries prediction. Effective and improving learning is observed with increasing system sizes of up to 108 qubits, demonstrating the largest quantum machine learning experiment to date. We further observe comparative quantum kernel advantage in learning tasks by constructing synthetic datasets based on the geometric differences between generated quantum and classical data kernels. Our findings demonstrate the potential of utilizing classically intractable quantum correlations for effective machine learning. We expect these results to stimulate further extensions to different quantum hardware and machine learning paradigms, including early fault-tolerant hardware and generative machine learning tasks.”
(Tuesday, July 09, 2024) “Quantum computing is an emerging field on the multidisciplinary interface between physics, engineering, and computer science with the potential to make a large impact on computational intelligence (CI). The aim of this paper is to introduce quantum approximate optimization methods to the CI community because of direct relevance to solving combinatorial problems. We introduce quantum computing and variational quantum algorithms (VQAs). VQAs are an effective method for the near-term implementation of quantum solutions on noisy intermediate-scale quantum (NISQ) devices with less reliable qubits and early-stage error correction. Then, we explain Farhi et al.’s quantum approximate optimization algorithm (Farhi’s QAOA, to prevent confusion). This VQA is generalized by Hadfield et al. to the quantum alternating operator ansatz (QAOA), which is a nature-inspired (particularly, adiabatic) quantum metaheuristic for approximately solving combinatorial optimization problems on gate-based quantum computers. We discuss connections of QAOA to relevant domains, such as computational learning theory and genetic algorithms, discussing current techniques and known results regarding hybrid quantum-classical intelligence systems. We present a schematic of how QAOA is constructed, and also discuss how CI techniques can be used to improve QAOA. We conclude with QAOA implementations for the well-known maximum cut, maximum bisection, and traveling salesperson problems, which can serve as templates for CI practitioners interested in using QAOA.”
(Wednesday, July 10, 2024) “The central challenge of quantum computing is implementing high-fidelity quantum gates at scale. However, many existing approaches to qubit control suffer from a scale-performance trade-off, impeding progress towards the creation of useful devices. Here, we present a vision for an electronically controlled trapped-ion quantum computer that alleviates this bottleneck. Our architecture utilizes shared current-carrying traces and local tuning electrodes in a microfabricated chip to perform quantum gates with low noise and crosstalk regardless of device size. To verify our approach, we experimentally demonstrate low-noise site-selective single- and two-qubit gates in a seven-zone ion trap that can control up to 10 qubits. We implement electronic single-qubit gates with 99.99916(7
Share this:
- Threads
-
Categories Quantum Computing Tags AI, algorithm, computer science, error correction, Futurum Group, machine learning, NISQ, physics, QAOA, quantum computing, qubit, space, trapped ion