Demonstration of qubit operations below a rigorous fault tolerance threshold with gate set tomography

Demonstration of qubit operations below a rigorous fault tolerance threshold with gate set tomography

Robin Blume-Kohout Center for Computing Research, Sandia National Laboratories, Albuquerque, NM 87185, USA    John King Gamble Center for Computing Research, Sandia National Laboratories, Albuquerque, NM 87185, USA    Erik Nielsen Sandia National Laboratories, Albuquerque, NM 87185, USA    Kenneth Rudinger Center for Computing Research, Sandia National Laboratories, Albuquerque, NM 87185, USA    Jonathan Mizrahi [ Sandia National Laboratories, Albuquerque, NM 87185, USA    Kevin Fortier Sandia National Laboratories, Albuquerque, NM 87185, USA    Peter Maunz Sandia National Laboratories, Albuquerque, NM 87185, USA
July 12, 2019

Quantum information processors promise fast algorithms for problems inaccessible to classical computers. But since qubits are noisy and error-prone, they will depend on fault-tolerant quantum error correction (FTQEC) to compute reliably. Quantum error correction can protect against general noise if – and only if – the error in each physical qubit operation is smaller than a certain threshold. The threshold for general errors is quantified by their diamond norm. Until now, qubits have been assessed primarily by randomized benchmarking, which reports a different “error rate” that is not sensitive to all errors, and cannot be compared directly to diamond norm thresholds. Here we use gate set tomography (GST) to completely characterize operations on a trapped-Yb-ion qubit and demonstrate with very high () confidence that they satisfy a rigorous threshold for FTQEC (diamond norm ).

Current Address:] Joint Quantum Institute, University of Maryland Department of Physics and National Institute of Standards and Technology, College Park, MD 20742

Figure 1: Gate set tomography is a robust procedure to characterize as-built qubits. (a): Gate set tomography (GST) models the qubit (a single Yb ion localized in a linear surface electrode ion trap) as a “black box” with a classical interface, and is agnostic to its physical details. (b): Flow chart of GST analysis. Its core is an iterative minimization, over data from increasingly long gate sequences, seeded with a linear inversion estimate. The final step, likelihood maximization, produces an accurate and unbiased estimate of the gate set. (c): A subset of the nearly 5000 data points taken: observed “bright” counts (black) for sequences of the form and the GST estimate’s prediction (red; see Fig. 3). Deviations from ideal gates appear only at . (d): GST achieves high precision from periodic sequences based on short “germs”. Here, the 11 germs used for this experiment are shown (colored boxes), as is the “fiducial sandwich” form of a general GST sequence.

I Introduction

The field of quantum information processing has seen great growth over the past thirty years, driven by exciting quantum algorithms inaccessible to classical computers. Small quantum information processors have been demonstrated experimentally using superconducting circuits nakamura:1999 (); wallraff:2004 (); barends:2014 (), electrons in semiconductors petta:2005 (); shulman:2012 (); veldhorst:2015 (), trapped atoms and ions leibfried:2003 (); barrett:2004 (); reiserer:2014 (), and photons obrien:2003 (); peruzzo:2010 (); politi:2009 (). Trapped ions are among the most reliable qubits available today; up to 14 qubits have been addressed in a single trap MonzPRL11 (), a 5-qubit quantum information processor has been realized Debnath2016 (), and single-qubit gates have demonstrated randomized benchmarking infidelities as low as harty:2014 (); BrownPRA11 (); Mount2015 ().

Unlike classical bits, qubits are intrinsically noisy and error-prone, and will require active, fault-tolerant quantum error correction (FTQEC QEC2013 ()) to operate reliably. To function, FTQEC requires physical qubit operations to be high quality, with errors below a specific threshold. Fault tolerance (FT) thresholds for quantum computing have been proven against various noise models, and generally require per-gate failure rates between and knill:2005 (); Aliferis:2006 (); aliferis:2007 (); Aliferis:2009 (). However, the particular metric of “error rate” depends on the noise model. Against realistic general errors, including small unitary errors, thresholds are stated in terms of the gates’ diamond norm error, AliferisNJP09 (); Aliferis:2006 (); Aharonov:2008 ().

Randomized benchmarking (RB) EmersonScience2007 (); Knill2008 (), the most commonly used method for qubit characterization, measures a single error rate () that closely approximates the gates’ average process infidelity. Because RB is relatively insensitive to unitary errors SandersNJP16 () that dominate diamond norm error Kueng15 () and have unpredictable consequences for FTQEC Aliferis:2006 (), it cannot efficiently measure diamond norm error to high precision. This makes it nearly impossible to demonstrate suitability for fault tolerance using RB alone, unless errors are assumed to be strictly incoherent. There are variants of RB that characterize and report additional parameters, but none of them are well suited for diamond norm characterization or comparison to fault tolerance thresholds MagesanPRL12 (); kimmel13 (); WallmanNJP15 ().

We use a characterization method called gate set tomography (GST) GST2013 (); Greenbaum15 (); GST2015 () to systematically debug and improve a 1-qubit trapped-Yb-ion quantum information processor, and – finally – to demonstrate with very high confidence that all three of its quantum logic operations surpass a proven threshold for FTQEC. GST provides a full and extremely accurate tomographic description of every gate, complete with statistical confidence bounds. We use this information to iteratively improve our single qubit operations and to place tight bounds on the diamond norm error of the final gates, producing the first single-qubit gates whose errors are demonstrably below a rigorous threshold for fault-tolerant error correction.

This is not a demonstration of FTQEC, which requires not just single-qubit gates, but also high-fidelity two-qubit gates, repeatable measurements, and (of course) more qubits. However, the GST methods that we use here to demonstrate 1-qubit gate errors below the threshold do generalize to 2-qubit gates, to the characterization of repeatable measurements, and to important properties of multiqubit systems such as crosstalk. So, while pushing single-qubit gate errors below the threshold is only one step toward achieving FTQEC, it is an important one.

Ii Results

ii.1 Gate set tomography

Our goals are (1) to implement quantum operations satisfying a FTQEC threshold, and (2) to “prove” – i.e., demonstrate conclusively – that we have done so. Genuine proofs are the domain of mathematics. In experimental science, the highest achievable standard is to provide experimental data (or summary statistics) that: (1) are consistent with the desired outcome; and (2) are inconsistent with any other plausible theory, and thus rule out all alternatives to some high level of confidence. Our intent is not to provide an exclusive protocol for such demonstrations, but rather to establish that gate set tomography is sufficient to do so.

While RB and quantum process tomography ChuangJMO97 (); WeinsteinJCP04 (); ObrienPRL04 () can be used in this fashion, they each face nontrivial obstacles. RB’s insensitivity to unitary errors makes it a poor tool for bounding worst-case error rates (diamond norms). In process tomography ShulmanScience12 (); MerkelPRA13 (), small calibration errors in the gates used to implement different measurements propagate to the final results, invalidating them. Gate set tomography (GST) is a self-calibrating tomography protocol that solves both of these problems. GST protocols based on short quantum circuits were developed at IBM MerkelPRA13 () and Sandia GST2013 (). The long-sequence GST protocol demonstrated here is orders of magnitude more precise.

GST relies on two assumptions: (a) the system being characterized is a qubit with a 2-dimensional Hilbert space; (b) each gate operation is stationary and Markovian. It treats the qubit as a black box with operation buttons (one for initialization, one for measurement, and the rest for gate operations) as shown in Fig. 1a, and self-consistently determines all operations up to a choice of basis (a gauge; see Methods). It can also detect and quantify violations of these assumptions (see next subsection).

In GST, the real (noisy) gates are modeled as trace-preserving linear maps on density matrices (TP maps). Such maps must be completely positive to be physical, and thus are usually referred to as “CPTP maps”; for technical reasons, we do not always impose the CP constraint in GST, but otherwise these maps are functionally the same as CPTP maps.

The qubit’s quantum state is a 4-element vector in the vector space of Hermitian matrices (Hilbert-Schmidt space) GST2013 (), and each gate is a matrix that acts on by left multiplication (i.e., ). Measurement is represented by a 2-outcome positive operator-valued measure (POVM) . Our target state preparation and measurement (SPAM) are and .

Data for GST come from gate sequences (quantum circuits), each comprising: (1) initialization, (2) a series of gates, and (3) measurement. Each sequence is repeated times, and the frequency of 0/1 counts is recorded. In the experiments reported here, we implemented and used the set of gates , but GST can analyze any gate set rich enough to prepare an informationally complete set of probe states and measurements.

GST analysis proceeds as shown in Fig. 1b. First, a specific set of short sequences is analyzed by linear inversion (see Methods) to get a rough estimate of the gates and SPAM operations. This estimate has an unavoidable gauge freedom; every observable probability is invariant under


for any invertible matrix . We choose a gauge that makes the estimated gates as similar to the target gates as possible (see Methods). If the rough estimate is not already completely positive, we truncate each gate to the nearest CP map, to ensure physically valid probabilities in the next step. Next, using the rough estimate as a starting point, we iteratively add more data. In the iteration, we add data from gate sequences of length into the pool, then numerically adjust the estimate to minimize the divergence between the observed frequencies and estimated probabilities. This “min-” estimate is then used as the seed for a numerical maximization of the likelihood function . Finally, we perform another gauge optimization to maximize similarity to targets.

The GST gate sequences (see Fig. 1d) are chosen to (collectively) amplify every physical parameter in the gate set. Short sequences called germs are repeated many times, and these “germ power” sequences are pre- and post-fixed by each of six fiducial sequences. In this work, we use six fiducial sequences, , where denotes the null sequence, and () are noisy rotations about (). These fiducials map and to (approximately) the six Pauli eigenstates, defining an informationally complete experimental reference frame. For further details on sequence design and a complete list of all experiments performed, see Methods.

ii.2 Experiment

Our qubit is a single Yb ion in a state-of-the-art linear surface ion trap (Fig. 1a). Ions are trapped by photoionizing neutral ytterbium vapor that reaches the trapping volume through a slot from the back of the surface trap chip. The qubit is encoded in the hyperfine clock states of the ground state of Yb: , . Standard laser cooling techniques are applied to Doppler cool the ion and prepare it in the state olmschenk_manipulation_2007 (). Standard fluorescence state detection olmschenk_manipulation_2007 () is used to measure in the basis.

Three logic gates – (the idle or identity gate), (a rotation), and (a rotation) – are realized by using a microwave horn to apply pulses near-resonant with the separation of the qubit levels. Broadband composite pulses (BB1) wimperis_broadband_1994 (); merrill_progress_2012 () are employed to minimize sensitivity to amplitude fluctuations in the microwave signal.

Figure 2: Progressive improvement of quantum operations. Over the course of 1 year, we used GST to improve our qubit, ending with clearly sub-threshold error rates. All metrics are computed using GST estimates based on data taken at the given time, but analyzed using best available algorithms at publication time. (a): Process infidelities of the three gates vs. wall time. (b): Diamond norm distance from estimated gates to targets, vs. wall time. Experiments from March 2015 surpass the best known diamond norm threshold of with 95% confidence, satisfying the threshold for fault-tolerance established in Aliferis:2009 (). (c): Violation of Markovian model (in standard deviations ) vs. wall time (see the section “Quantifying non-Markovianity” for details). Non-Markovian noise was progressively eliminated (e.g., by adding drift control and dynamical correction; see main text), guided by GST.
Figure 3: Process matrices and error generators for the final gates implemented 30 March 2015. (a): GST estimates of the , and gates, shown as superoperators in the basis of Pauli matrices, and based on data from gate sequences out to length 8192. For each estimate , we also show and to emphasize errors. Bar height shows absolute value of matrix elements. White bars are fixed by the TP (trace-preserving) constraint, red ones should (ideally) be , blue ones should be , and beige/teal ones should equal but are positive/negative, respectively. Wireframes indicate the ideal (target) gates for comparison. (b): Error generators for each gate, using the same colors as (a). We define these as , where is the estimate and is the target.

Using characterization procedures to debug and improve qubits has a long history. A Ramsey fringe or Rabi oscillation experiment is a kind of limited tomography, combined with physical intuition, which is used to improve the quality of the quantum logic operations. In typical tune-up procedures, different types of these experiments are iterated, until the qubit gates are deemed good enough to proceed.

Various improvements to this generic tuning-up scheme outlined above have been offered, including augmenting oscillation experiments to detect microwave pulse distortions Gustavsson:2013 (), designing small sequences for error amplification Morton2005 (), supplementing sequence experiments with RB to do detailed noise spectroscopy OMalley:2015 (), or replacing them with iterative randomized benchmarking to guide the system toward higher RB fidelity operations Kelly:2014 (). All of these techniques, and other experiments combined with physical intuition, can be and have been used to produce qubits with very high-fidelity operations. In comparison, GST has the distinct advantage in that it includes all experiments necessary for full and loophole-free qubit characterization. It can be though of as systematic statistical inference on a provably sufficient set of Rabi/Ramsey experiments.

We used GST to analyze systematically and improve our trapped ion qubit operations over the course of five experimental runs from 17 April 2014 - 30 March 2015. Experiments #1-2 used the Sandia Thunderbird trap stick_demonstration_2010 (), and Experiments #3-5 used Sandia’s high-optical-access (HOA-2) trap. Fig. 2a summarizes the gates’ steady improvement over this period by tracking their process infidelities MagesanPRA11 (), which corresponds to the RB error rate MagesanPRA2012 ().

Experiment #1 detected severely non-Markovian behavior. We sought to address this by stabilizing the microwave amplifier’s temperature, and stabilizing microwave -times using active feedback (drift control), as described in Methods. Experiment #2 showed improved fidelity in the and gates, but no reduction in non-Markovianity. We then moved our qubit to the HOA-2 trap, and improved trap stability. In Experiment #3, GST showed significant improvements in fidelity and Markovianity, and that remained worse than the other gates. To improve it, we changed from “do nothing for one clock cycle” to the dynamical decoupling pulse sequence , where and denote rotations around and , respectively, and means “wait for the duration of a -pulse”. We also applied active drift control of the qubit frequency, and improved the calibration of the BB1 pulse sequences. Experiment #4 showed reduction of coherent errors in and , but persistent non-Markovian errors in . After we upgraded to the 2nd-order dynamical decoupling sequence  khodjasteh_dynamical_2009 (), Experiment #5 demonstrated uniformly excellent gates. Subsequent analysis indicates that the improved performance of stemmed largely from the constant duty cycle of the microwave system, rather than from the intrinsic properties of the decoupling sequence used. This emphasizes that GST can identify specific errors, but not necessarily their cause. The estimated process matrices for the gates are shown in Fig. 3a. Fig. 3b shows the error generators, defined as , where is the estimate and is the target.

ii.3 Demonstrating suitability for fault tolerance

Useful quantum computation is expected to require fault tolerant error correction. The most important milestone for a quantum operation is, therefore, “Is it suitable for use in FTQEC?” Operations that induce too much error will cause FTQEC protocols to fail. Demonstrating conclusively that gates are suitable for fault tolerance requires: (1) establishing a sufficient condition for the gates to not induce failure; and (2) showing that the gates satisfy that condition, with high confidence, by means of experimental data that are inconsistent with all gates that don’t satisfy the condition.

Demonstrating suitability for fault tolerance using infidelity alone is hard. Threshold theorems against general errors (arbitrary CP maps) are stated in terms of the diamond norm distance between the real and ideal gates aharonov:1998 (); Aharonov:2008 (),


where is the system’s Hilbert space dimension, is the trace norm, and the supremum is over density matrices with dimension Benenti2010 (). Because the diamond norm error can be as large as Kueng15 (); SandersNJP16 (), even a spectacular RB result like harty:2014 () only establishes an upper bound of on the diamond norm. The best-known proof of fault tolerance against general noise Aliferis:2006 () derived a threshold of against stochastic noise, and generalized it to a diamond norm threshold of against general (coherent) noise. This was subsequently improved to aliferis:2007 () and finally to Aliferis:2009 (), the highest (currently) proven threshold against general noise.

Unlike RB, GST enables direct computation of the diamond norm between the estimated and target gates (we use a semidefinite program Watrous2012 ()). Fig. 2b shows the diamond norm error of our gates over time, culminating on 30 March 2015 in diamond norm error rates (with 95% confidence intervals) of , , and for , , and respectively. All three gates surpass the threshold with 95% confidence. (In point of fact, they surpass even the older threshold with 95% confidence.)

We note that, although we only demonstrated Clifford operations, and non-Clifford operations are needed for universal control, FTQEC is possible only using Cliffords. Furthermore, we can still extrapolate the performance of non-Clifford gates (e.g., a T gate) in our system. A pessimistic estimate of the error on an rotation, for example, would simply be the same as the error on the gate that we characterized. This is because implementing the gate in practice is equivalent to running the gate for a shorter duration.

ii.4 Quantifying non-Markovianity

Figure 4: box plots for gate set tomography fits for three datasets. In each individual box, is summed over a set of 36 different gate strings. If the underlying system is Markovian, is (approximately) a random variable. The box color indicates the score for that collection of sequences. Grey indicates score values that are expected due to statistical fluctuations, while red indicates significant model violation at 95% confidence. That is, if the gates are Markovian, the probability of observing even one red square is at most 5%. Note that certain germs are too long to appear at . (a): box plot for simulated Markovian data. (b): box plot for experimental data from 2 March, 2015, with indications of strong non-Markovianity. (c): box plot for experimental data from 30 March, 2015, with vastly decreased non-Markovianity.

In real experimental systems, repeated quantum operations are never actually identical. For example, experimental imperfections in the stability of the system may cause quantum operations to drift over time. Collectively, we refer to all such non-repeatability as non-Markovianity. It represents a significant potential problem for fault tolerance, as proofs of fault tolerance thresholds are typically carried out using Markovian error models. So, to be confident that a gate set is suitable for FTQEC (i.e., achieves a fault tolerance threshold), we would like to demonstrate that non-Markovian behavior is absent. This is not feasible, for two reasons. First, all physical systems (including qubits) are at least a tiny bit non-Markovian. Second, “non-Markovian noise” is so general that there is always some conceivable mechanism that would elude detection by any protocol (not just GST). Our goal is to reduce detectable non-Markovian behavior to the point where its visible effects are consistent with the FT threshold.

We use GST results to debug non-Markovian effects and achieve this goal, as illustrated in Fig. 2c. Doing this is nontrivial, because neither GST nor process tomography is actually designed to characterize non-Markovianity. In GST’s underlying model, the qubit is Markovian: its state at time is determined completely by (1) its state at time and (2) the operation applied at time . This assumption is far reaching. It implies that noise in the logic gates is stationary, uncorrelated in time, memoryless, and independent of context (e.g., what gates were recently applied). It implies that the gate operations can (for a single qubit) be represented as static superoperators, and that state preparation and measurements may each be represented as static four-dimensional vectors and dual vectors (respectively), in Hilbert-Schmidt space.

As indicated above, the Markovian assumption is not strictly true for any experimental system. In addition to slow drift, there may be correlations between errors in consecutive gates, and the “qubit” may not even be a two level system (e.g., due to leakage levels). These are all examples of non-Markovianity, and lie outside the GST model.

In principle, all guarantees about GST are void in the presence of non-Markovian noise, as there are no process matrices to measure or report. However, for many typical non-Markovian behaviors, GST degrades in a quantifiable way. These kinds of non-Markovian noise cause data that are consistent with no Markovian gate set, and this failure to fit the data can be quantified. Since data generated by any Markovian model could be fit with predictable accuracy, significant badness-of-fit can be interpreted as violation of the model and therefore as non-Markovianity, though the particular type cannot be easily identified. As long as the data appear sufficiently Markovian, the GST estimate will be fairly reliable and have significant predictive power.

To quantify non-Markovianity, we consider the log-likelihood (Eq. 13)


where The best conceivable fit to a dataset would be one where for every sequence . Thus, the entropy of a dataset is an upper bound on ,


We define the quantity .

Standard properties of maximum likelihood estimation theory WassermanBook13 () imply that if

  1. the data were in fact generated by some gate set,

  2. there are free parameters in the gate set, and

  3. the dataset contains distinct gate sequences

then as , is a random variable with a distribution, where This means that its expected value is , and its RMS variance is . Thus, if the fit is “good”, then should lie roughly within the interval . Hence, by comparing the difference to we can determine how well the Markovian model was able to fit the data.

We quantify goodness-of-fit by , the number of standard deviations from the expected mean the expected mean the log-likelihood score is:


We can also calculate for individual experiments or subsets of gate sequences. Figure 4 illustrates this, where is shown for every individual experiment associated with each power of each germ (for a total 36 experiments per collection, due to six preparation fiducials and six measurement fiducials). This analysis makes it possible to see whether non-Markovianity increases with sequence length (it usually does, because longer sequences amplify slowly varying noise), and which sequences are particularly inconsistent with the best Markovian fit.

Figure 4 compares scores for a simulated (perfectly Markovian) dataset to those for two experimental data sets, one from 02 March 2015, and the other from 30 March 2015. The March 2 experimental dataset is highly non-Markovian, while the March 30 dataset looks very similar to the simulated dataset. These data demonstrate the degree to which we are able to stabilize our qubit and reduce non-Markovian effects. In the final run on March 30, Markovianity is violated only at the level. While this is statistically significant – it implies with high confidence that the gates are not perfectly Markovian – it is not practically significant. To see this, recall that this is an extraordinary sensitive experiment, as witnessed by the fact that the error bars on the diamond norm are . This sensitivity extends to non-Markovian behavior as well. Reducing the sensitivity of the experiment by a factor of 4 (either by reducing maximum by a factor of 4, or by reducing by a factor of 16) would render the non-Markovianity undetectable, at the cost of increasing the error bars by a factor of 4 to . This implies that the observed non-Markovianity is effectively equivalent to less than additional diamond norm error, which is comfortably below the threshold.

ii.5 Comparison to randomized benchmarking

As of this writing, randomized benchmarking (RB) is the de facto standard in qubit characterization. As a consistency check, we perform RB simultaneously with the final GST experiment (by interleaving the GST and RB sequences over the entire period of experimentation), to see whether GST correctly predicted the results of RB. We follow the experimental and analysis procedure of Ref. wallman_randomized_2014 (), and use RB sequences ranging in length from 2 gates to 1970 gates.

Strictly speaking, RB measures the error rate per Clifford operation. Our Clifford operations are, as is usual, compiled into elementary gates, with an average of elementary gates per Clifford. Analysis of the data in strict accordance with the literature (i.e., plotting survival probability versus # of Cliffords in the sequence) yielded an experimental error rate of per Clifford operation. Dividing this by 3.125 (a questionable but common practice) suggests a per-gate error rate of about .

However, our main goal is to compare the RB data with GST’s predictions for it. For this purpose, we find it more informative to fit (and plot; see Fig. 5) the observed probabilities versus the number of elementary gates in the sequence. All the rest of the analysis in this section is based on this analysis method, which yields a per-elementary gate RB error rate of . Error bars are 95% confidence intervals. The experimental error bars are calculated via non-parametric bootstrap (by resampling the experimental data with replacement). We then simulate those RB experiments using the GST estimates. The GST results predict an RB error rate of (see Figs. 5a and 5b). The simulated error bars are calculated via parametric bootstrap. (The GST estimate is used to generate many sets of simulated GST experiments, each of which in turn yields a new GST gate set estimate. This ensemble of estimates then generates an ensemble of simulated RB decay rates, from which the simulated error bars are derived.)

While these decay rates are nearly identical, there is a statistically significant discrepancy. The most obvious explanation is a flaw in the GST analysis, but we find that extensive simulations with known Markovian gates rule this possibility out. We believe that the discrepancy stems from physical causes – i.e., from non-Markovian noise. The most common form of non-Markovian errors is low-frequency drift, which manifests in both RB and GST as coherent errors that remain nearly fixed over the course of any one sequence, but change from sequence to sequence (and between repetitions of a single sequence). In the presence of such effects, GST typically overestimates the RB decay rate, because GST amplifies coherent errors to which RB sequences are relatively insensitive (Markovian or not). Thus, GST typically reports a higher rate of Markovian noise in a quixotic attempt to fit its data, while RB simply doesn’t see the noise.

But in this experiment, we observe the opposite effect. Instead of overestimating the RB error rate, GST underestimates it. While the exact cause remains uncertain, we observe that this behavior is completely consistent with anti-correlated noise (each gate flips between under- and over-rotation at each application) induced by dynamically corrected gates khodjasteh_dynamical_2009 ().

Here is a concrete model that reproduces this behavior: Consider a unitary error that varies in time – but instead of varying slowly, it oscillates at the system’s Nyquist frequency (i.e., flips sign every clock cycle). For simple gates implemented with a single pulse, this would be highly implausible. In this experiment, however, we implement dynamically corrected gates (DCG). The simplest DCG is a dynamically corrected idle gate (our ). This is nothing but dynamical decoupling – periodic pulses that echo away small rotations. Such sequences create a “toggling frame” for the qubit that flips sign twice per clock cycle. Any timing or amplitude errors in the pulses can leave a residual error that flips sign every clock cycle, making a plausible noise model for a DCG.

We model this effect by augmenting the qubit state space with a classical binary variable . We define a composite gate set , based on a standard gate set , which consists of two single-qubit gate sets and that act conditionally on the value of the classical bit , which flips every time a gate is applied. These gate sets are identical to , except that the and elements of have a fixed, slight over-rotation by an angle , while the and elements of have a fixed, slight under-rotation by . At the beginning of each simulated experiment, is chosen randomly.

acts on an -dimensional state space, and data generated from it is not fully consistent with any Markovian single-qubit gate set. But GST can be applied to that data, and will find the Markovian single-qubit gate set that fits it best. (Indeed, as no experimental system is perfectly Markovian, this is in essence what GST always does.)

We generate simulated data, with finite sample error, for all the GST and RB experiments performed on 30 March, 2015. For this simulation, is defined by setting equal to the GST estimate from 30 March, 2015, and setting .

Analyzing the GST data generated by yields an estimated gate set almost identical to that obtained from experimental data (). All but two of the 36 free gate matrix elements are within the 95% confidence intervals assigned to , and the three remaining elements are at most outside them. Every gate is within (in diamond norm) from the corresponding gate in . We conclude that GST, as performed, cannot distinguish the composite model () from .

Figure 5: Randomized benchmarking results. Colored dots are experimental or simulated data points; lines are exponential decay fits to the data. (a): Experimental RB data. (b): RB data simulated using the gate set derived from experimental GST results. (c): RB data simulated using the non-Markovian gate set . Here, is generated using the composite Nyquist-limited noisy gate set model proposed in the section “Comparison to Randomized Benchmarking.” This model toggles between slightly over- and under-rotated gates with every gate application, which is a reasonable scenario for our qubit due to our use of dynamically corrected gates. The experimental RB decay rate is , which is indistinguishable from ’s RB decay rate of , but distinct from ’s RB decay rate of , demonstrating the plausibility of our non-Markovian model in explaining the apparent discrepancy between experimental RB and GST.

The RB data simulated with also matches the experimental RB data almost perfectly, yielding an RB error rate of that is statistically indistinguishable from the experimentally observed RB decay rate of . Both datasets (experimental and simulated-by-) are shown in Figure 5, along with RB data simulated from . We conclude that RB observes significantly different error rates for and .

This does not imply that our qubit really is described by , but it demonstrate a plausible non-Markovian model that is fully consistent with our data. There might be many other (different) non-Markovian models equally consistent with it. And while certain kinds of non-Markovian noise can be detected by RB Ball2016 () and GST, neither GST nor RB are designed to function reliably in the presence of any non-Markovian noise, so neither of them is explicitly “right” or “wrong” for this case.

ii.6 The relative power of RB and GST

RB and GST share the common framework of data from gate sequences (circuits) that are (1) diverse, (2) repeated, and (3) long. But they are distinguished by the kind of sequences performed. RB sequences are random, for the specific purpose of “twirling” the noise. GST sequences are structured and periodic, for the specific purpose of amplifying errors.

This difference is fundamental. It makes RB intrinsically insensitive to coherent errors, which dominate the diamond norm error metric SandersNJP16 (); Kueng15 (). For example, suppose that one logic gate over-rotates by a small angle , while the others are perfect. In random sequences containing applications of this gate, it will (by construction) be interleaved with other gates chosen randomly. The rotations by will add up incoherently, producing (on average) a total rotation of , and therefore an error probability of . Thus, a coherent error by appears (in RB) as an incoherent error of probability .

But circuits of practical interest are not random. Since not all “useful” circuits are known at this time, it is wise to consider how errors affect arbitrary circuits in the worst (most fragile) case. For the example given above, the worst case is a periodic sequence in which the imperfect gate appears consecutive times. Rotations add up coherently, the final angle is , and the final error probability is . So, for example, a rotation can cause a failure rate after just gates. In randomized circuits, the same failure rate would require gates.

The diamond norm metric is a strict upper bound on the rate at which failure probabilities can grow, and so it takes account (by construction) of the worst-case behavior given above. The diamond norm error for a small coherent error by angle is . Process infidelity (closely related to the RB error rate) does not account for worst-case behavior, and the process infidelity for a small coherent error by angle is .

GST intentionally implements a wide variety of periodic sequences, to ensure that at least one of them is approximately “worst case” for every possible coherent error. This allows GST to detect coherent errors of size using sequences of length , repeated times. Detecting the same error with randomized sequences requires much long sequences of length , or else a much higher number of repetitions (both of which correspond to orders of magnitude more time and effort).

Periodic sequences might be incorporated into RB, to make it more sensitive. Doing so, however, would eliminate its characteristic feature. Such a protocol would no longer be randomized benchmarking. On the other hand, there are several interesting variations of RB that retain its randomized nature, most notably interleaved benchmarking MagesanPRL12 (), RB tomography kimmel13 (), and unitarity benchmarking (URB) WallmanNJP15 (). While interesting in their own right, they are all subject to the same trouble: random gate sequences are much less sensitive to coherent errors than periodic ones, and therefore every form of RB is necessarily inefficient at detecting coherent errors.

Unitarity benchmarking is particularly interesting, since (unlike other forms of RB) it can separate coherent and incoherent errors, and therefore provide good information about diamond norm error rates. Unfortunately, it is (compared with GST) extremely inefficient at doing so.

Wallman et al. WallmanNJP15 () defined a quantity (unitarity), which measures the rate of purity decay. They gave an RB-like protocol for measuring it, and pointed out that and together could be used to bound the diamond norm error. If


then the noise is purely incoherent, and the diamond norm error is . If is sufficiently small, then the errors are primarily incoherent, and the diamond norm error remains .

However, the actual bounds (see Wallman15FT ()) are of the form


We have demonstrated that our gates’ diamond norm error is using GST. Doing the same thing using unitarity requires showing that . But is itself an RB-type quantity, meaning that it appears as an error rate (in experiments that measure purity), and is measured using RB. As a result, showing that is equivalent to:

  1. Performing standard RB to measure , the decay rate of sequence fidelity.

  2. Performing a different RB-like experiment to measure .

  3. Demonstrating (based on those experiments) that .

For (the regime we access experimentally), this requires measuring both and to precision. This is extraordinarily hard. The most efficient way to do it is using sequences of length . These would yield survival probabilities around . Achieving the necessary precision would require estimating to , which would require approximately repetitions (because the uncertainty is ). This is at least times more repetitions than would be required for standard RB, or for GST, and is completely impractical.

ii.7 Validating accuracy with simulations

We have claimed uncertainties (error bars / confidence regions — see Methods) of about for diamond norms and process matrix elements. This is remarkable, and demands supporting evidence. To confirm this behavior we simulate GST experiments using (known) gate sets with unitary errors. The results (Fig. 6) confirm Heisenberg scaling: diamond norm distance between estimated and true gates decreases with the maximum sequence length () as . This scaling holds up to , where is the stochastic error rate. This is consistent with the observed error bars on diamond norm errors in our final experiment, for which .

Figure 6: Confirmation of error scaling in gate set tomography. Here, we show diamond norm difference between true and estimated gates in simulated GST with small unitary errors. Mean diamond norms are shown, averaged over 100 trials. Estimation error scales as , where is the maximum sequence length in the data. Each trial uses samples per experiment.

Iii Discussion

Gate set tomography allows us to achieve high-quality gates in a trapped-Yb-ion qubit, and to characterize it to unprecedented precision. Although lower RB error rates have been reported in trapped-ion qubits BrownPRA11 (); harty:2014 (), our gates are the first to demonstrably surpass a rigorous fault tolerance threshold against general noise. GST is the first protocol that can efficiently demonstrate this important milestone and provide reliable feedback to debug and improve those gates.

Low-error single-qubit gates are just one of several critical achievements required to enable fault tolerant quantum computing. Thus, this is only a first step. But GST – which can be generalized to 2-qubit gates and measurements – does answer one key and pressing question: “Once suitable operations have been achieved, how can their performance be verified for a critical, objective observer?” Randomized benchmarking can provide reliable information about process fidelity (which unambiguously captures stochastic or incoherent errors), but as of this writing, process fidelity is not known to be the relevant metric for fault tolerance. An exciting recent development in this area is the introduction of randomized compiling Wallman15n (), which has the potential to provably reduce the importance of coherent errors. But until and unless such techniques lead to a fault tolerance proof that is insensitive to them, and are confirmed to be practical in the context of FTQEC, coherent errors remain a point of concern. GST provides an efficient way to diagnose and bound all Markovian errors in gates.

Iv Methods

iv.1 Experimental details

In Sandia’s Thunderbird trap, ions were trapped above the trap surface. Typical trap frequencies were , , and , for the axial and two radial modes, respectively. In the HOA-2 trap, ions were trapped above the trap surface and trap frequencies of , , and were achieved. Typical trapping times were several hours for the Thunderbird trap and up to for the HOA-2 trap. Coherence times were measured to be in both traps, and were most likely clock-limited.

The microwave radiation used for qubit manipulation was generated by single-sideband (SSB) modulating the output of a dielectric resonator oscillator (DRO) with the output of a direct digital synthesizer (DDS) near . The master clock for the DRO and DDS was generated by a rubidium frequency standard. The output of the SSB modulator was amplified and directed parallel to the trap surface using a microwave horn. The microwave frequency and phase was controlled via the DDS and approximately square pulses were generated by switching the output of the DDS using a high-isolation rf switch. An offset was added to the constituent pulses of the BB1 pulse sequence to compensate for switching imperfections.

Drift control of the microwave -time was realized by interleaving experiments in which the ion was initialized, exposed to a microwave pulse, and measured. The -time was adjusted after state detection; upon seeing , -time was decreased by , while upon seeing , -time was increased by . For the next experiment, the -time was then truncated to the time resolution of the experimental control (. Similarly, drift control of the qubit frequency was implemented by interleaving a Ramsey experiment in which the ion, after state initialization, is subject to: (1) a gate, (2) a wait time, and (3) a gate. Upon state detection, the qubit frequency was adjusted by for a result, and by for a result.

iv.2 Linear GST

Linear-inversion GST (LGST) is a highly reliable but low-accuracy way to obtain an initial estimate of the gate set that serves as a seed for further refinement by long-sequence GST (see next subsection). LGST is essentially simultaneous “uncalibrated” process and state tomography. By performing process tomography-like experiments on a set of gates, as well as the null operation (i.e., the “do nothing for no time” operation), LGST can provide rough estimates of all the gates involved, as well as the state preparation and measurement operations. LGST requires minimal assumptions about the various operations (unlike standard tomography), and computes its estimates using only basic linear algebra (the most complicated step is matrix inversion). A detailed explanation of the LGST procedure is provided in GST2013 (); LGST is also described in Greenbaum15 ().

iv.3 Analyzing long sequences in GST

GST incorporates data from long sequences in two stages. The first stage consists of several iterations, each of which performs a minimum- estimation. Each iteration takes the result of the previous iteration as a seed, and includes successively more of the long-sequence data. The second stage is a maximum-likelihood estimation which is seeded from the first stage and uses all of the data. This procedure consistently avoids local minima in the objective function. In this section, we give the details of this algorithm (outlined in Fig. 1).

The iterative fitting procedure starts by fitting only data from the shortest gate sequences (which are easy to fit and insensitive to most non-Markovian noise), then successively adds longer and longer sequences (with base sequence length ). Since we get an estimate at each intermediate , it is possible to quantify not just the goodness of the best fit, but how the goodness-of-fit behaves as longer and longer sequences are added in, which is useful for debugging.

At each step in the iterative process, we vary the gate set to minimize Pearson’s test statistic, which measures the discrepancy between a predicted probability and an observed frequency . It is defined as


where is the number of samples taken. In this analysis, is used to compare the set of probabilities predicted by a gate set () and the frequencies obtained from a dataset (). Each experiment (i.e. gate sequence) is associated to two probabilities: “plus” has probability and “minus” has probability . The of a single gate string is


where is the number of times the experiment was performed, is the probability of a “plus” outcome as predicted by the gate set, and is the observed frequency of “plus”. The total for a dataset is just the sum


To estimate our gate set parameters, we minimize at each iteration using the Levenberg-Marquardt algorithim implemented in SciPy scipy ().

The final stage in long-sequence GST analysis is a maximum-likelihood estimation (MLE), based on numerical optimization of the log-likelihood function . The log-likelihood for an -outcome system with predicted probabilities and observed frequencies () is given by:


where is the total number of counts. Like the statistic, is used to compare the set of probabilities predicted by a gate set () to the frequencies obtained from a dataset (). Each experiment (i.e. gate sequence) is associated to two probabilities: “plus” has probability and “minus” has probability . The contribution of a single gate string is


where is the number of times the experiment was performed, is the probability of a “plus” outcome as predicted by the gate set, and is the observed frequency of “plus”. The total log-likelihood for an entire dataset is just the sum


We find the maximum of this quantity using the same Levenberg-Marquardt algorithm as above, in order to compute the final (modulo gauge optimization) estimate of the gates.

LGST would be a perfect estimator in the absence of finite-sample error. However, it is inefficient with respect to accuracy. Like process tomography, its inaccuracy scales as , which means that achieving error bars on all parameters would require around repetitions of each experiment. Long sequences amplify errors proportional to , enabling inaccuracy of for all parameters. (This scaling breaks down for , where is the rate of stochastic decoherence. In our experiments, , and we perform experiments as long as .)

We use a hybrid algorithm (involving both min- and MLE) because each of its components have certain weaknesses. Empirically, we find that MLE is statistically well-motivated and avoids any bias, whereas optimization is numerically more stable and faster computationally but yields biased estimators, especially for the SPAM parameters. Our hybrid method combines both virtues, by using the more efficient and reliable min- algorithm to get a very good seed for the final (unbiased) MLE.

On a modern laptop, single-qubit GST with maximum can run in under 1 minute; the analysis for maximum takes about 40 minutes.

iv.4 Selecting gate sequences for GST

The data that GST use to reconstruct a gate set come from performing gate sequences (i.e. quantum circuits). Every gate sequence necessarily comprises (i) initialization, (ii) some gates, and (iii) measurement (which yields a count that is recorded in the dataset). The sequences used for GST have an additional structure:

  1. Each GST sequence begins with a preparation fiducial sequence, and ends with a measurement fiducial sequence, with an “operation of interest” sandwiched in the middle.

  2. The “operation of interest”, which could in principle be any gate sequence, is chosen to be a germ power sequence – i.e., a short “germ” sequence, repeated an integer number of times.

Thus, every GST sequence is of the form , where and are preparation and measurement fiducials (respectively), is a germ, and is an integer. and range exhaustively over a set of 6 fiducial sequences, while ranges exhaustively over a set of 11 germs. In this section, we explain how the fiducials and germs are chosen.

The purpose of the fiducials is to prepare a sufficiently diverse set of input states and measurements to completely probe the operation of interest. This is achieved if (and only if) the input states and the measurement effects are both informationally complete (IC). A set of matrices is IC if and only if it spans the vector space of matrices. This requires at least linearly independent elements.

In general, any randomly chosen set of states or effects will be IC. So, for single-qubit GST, we could choose random fiducial sequences. However, while the resulting and will almost certainly be linearly independent, they may be close to linearly dependent. This property is quantified by the spectrum of the Gram matrix , defined by


If either set fails to be IC, the Gram matrix will fail to have non-zero (to machine precision) singular values. As any one of the largest singular values becomes small, inverting the Gram matrix on its support (as is required for LGST) becomes ill-conditioned, and finite-sample fluctuations in GST get amplified, causing poor accuracy.

We would like both preparation and measurement fiducials to be uniformly IC, meaning that they span as uniformly as possible, and the smallest singular value of the Gram matrix is as large as possible. There exists a single-qubit uniformly IC set with only 4 elements (the SIC-POVM), but it cannot be generated with Clifford operations and stabilizer states. The smallest convenient uniformly IC set is the 6-element set of stabilizer states (the eigenstates of , , and ). We choose six fiducial sequences so that, if the gates are ideal, they will prepare the stabilizer states exactly. They are


where indicates the null sequence (no gates).

Slightly imperfect gates will prepare states (and effects) that are close to the stabilizer states – and therefore close to uniformly IC, and almost as effective in probing the operation of interest. If the gates are sufficiently far from the targets, it can be detected by computing the singular values of the empirical Gram matrix, and then new fiducials can be chosen.

Once the fiducials are defined, we need to define “operations of interest” for them to probe. By sandwiching any such operation between an exhaustive set of 36 fiducial pairs, we are essentially doing process tomography on the operation (although the algorithm for incorporating these data into the GST fit is more complex than simple process tomography).

The obvious operations of interest are the gates themselves (, , and ). By probing each gate tomographically, and repeating each sequence times, GST can estimate the gates to with accuracy (for some constant ). To achieve higher accuracy, we do tomography on powers of the gates, by designating, e.g., or as an operation of interest. (Powers of 2 are chosen merely for convenience; any logarithmically spaced sequence of integer powers would work).

Repeating a gate times – i.e., performing sequences of the form – amplifies errors in the gate itself. So, e.g., if is actually a rotation by , then is a rotation by . GST can now characterize that rotation to within , which equates to estimating to within . Raising gates to the th power amplifies deviations by , which in turn reduces estimation error by a factor of .

However, simple repetition of does not amplify every error. For example, suppose that is in fact a rotation, but around the wrong axis, corresponding to the unitary map


as opposed to the target unitary ( and indicate the Pauli operators and ).

This is a tilt error, and it is not amplified by . It’s easy to see this by observing that , so the error cancels itself out after just 4 repetitions.

More sophisticated sequences are needed to amplify tilt errors. For this example, it is sufficient to probe . Assuming (for now) that is a perfect rotation around , is a rotation by . Therefore, performing amplifies the deviation by a factor of , and setting it as an operation of interest allows GST to estimate to within . The short sequence is a germ, and repeating it times yields a germ power sequence that can be sandwiched between fiducials to equip GST with high sensitivity to the parameter .

The general situation gets rapidly complicated – e.g., if is not perfect, then alone cannot distinguish between tilt in and tilt in . Each germ is sensitive to some nontrivial linear combination of gate set parameters. To choose a set of germs, we list the possible germs (i.e., all reasonably short sequences), and for each germ we identify what linear combination of parameters it amplifies. We do this by computing a Jacobian,


where is the gate sequence product for germ (obtained by just multiplying together the process matrices), and is a vector containing all the parameters of the gate set (e.g., the elements of all the process matrices).

In the single-qubit case, is a matrix, and is 48-dimensional because it contains the elements of three gate matrices. Constraining all gates to be trace-preserving reduces the number of free parameters to 12 and 36 (respectively), so is a matrix. Its 12 right singular vectors indicate linear combinations of gate set parameters that amplifies (when repeated times), and the corresponding singular values quantify how much they are amplified. A zero singular value indicates a parameter that is not amplified at all (like the tilt error discussed above). A set of germs is, collectively, described by a Jacobian


Our goal is to choose germs that provide high sensitivity at “large” values of . In practice, it is not useful to make larger than , where is the rate of stochastic or depolarizing noise. To select germs, however, we ignore this effect and make the simplifying assumption that the gates (and therefore ) are reversible (a good approximation when is small). Under this assumption, it is possible to define the limit of the Jacobian in Eq. 18. Using the product rule, and assuming that all the gates are unitary (and therefore ),


As , the average over all powers of twirls . By Schur’s lemma, the effect of twirling is to project onto the commutant of – i.e., onto the subspace of matrices that commute with . Furthermore, multiplication by the unitary is merely a change of basis, and has no effect on the right singular vectors or the singular values of . So, up to an irrelevant change of basis:


where is the projection onto the commutant of .

This framework defines a notion of informational completeness for germs. A set of germs is amplificationally complete (AC) if and only if the right singular rank of its Jacobian equals the total number of physically accessible (gauge-invariant) parameters in the gate set. For a general set of 3 single-qubit trace-preserving gates, a gauge transformation is where is an invertible trace-preserving superoperator, so there are 12 gauge parameters, and 36-12=24 gauge-invariant parameters. To build an AC set of germs, it is sufficient to add germs to the set until its Jacobian has rank 24. By constructing a complete set of infinitesimal gauge transformations, we can actually construct the projector onto the (local) space of gauge-invariant perturbations to the gate set .

We then optimize this set numerically, by adding and removing germs (taken from an exhaustive list of all short sequences), and only keeping a modification if it lowers a certain score function. (For single-qubit GST, we find it convenient to search over all germs of length . However, this set of candidates need not be exhaustive. (A larger gate set, for example, would generate a prohibitively large exhaustive candidate set.) We have used randomly chosen subsets as candidate sets and gotten similar results.) The score function is


This score estimates the mean squared error of estimation if a fixed number of counts are spread over the distinct germs. Running this algorithm until it cannot improve the germ set any further produces the following set of 11 germs used in the final (March, 2015) GST runs (see Fig. 7):

Figure 7: Sensitivity analysis for germ selection. Here, we show the sorted singular values of the Jacobian matrix for different germ sets for the standard gate set. Each singular value of the Jacobian corresponds to a gate set parameter; a large singular value indicates that the chosen germ set provides GST with sensitivity to that parameter. Given that there are always experimentally inaccessible gauge parameters, it is impossible to be sensitive to all parameters. The dashed red line indicates the number of gauge parameters for this gate set (14). The blue triangles are singular values for a Jacobian with the amplificationally complete 11-germ set used for the March 2015 GST runs; all singular values corresponding to non-gauge parameters are large. The purple triangles are singular values for the 3-germ set containing just the bare gates , , and . It is amplificationally incomplete, as indicated by the presence of near-zero singular values that correspond to non-gauge parameters.

iv.5 The GST gauge, and how to set it

A gate set comprises: an initial density matrix (represented as a Hilbert-Schmidt vector), a measurement effect (represented as a Hilbert-Schmidt dual vector), and one or more gates (represented as superoperators). But not every parameter in this representation is physically observable. A gate set has intrinsic “gauge” degrees of freedom, because two distinct gate sets (or an entire manifold of them) can yield identical probabilities for all possible experiments. Gauge transformations alter a gate set’s elements without changing any observable probability. They take the form


where is any invertible superoperator. If (as usual) we consider only trace-preserving (TP) gate sets, then the corresponding necessary and sufficient condition for a gauge transformation to preserve this constraint is that be itself TP (i.e., its first row should be .

This gauge freedom makes it difficult to compare two gate sets, since two apparently-distinct gate sets may actually be equivalent. Most of the metrics used to measure distance between two gates are not gauge-invariant (e.g. fidelity, trace-norm distance, and diamond-norm distance are all gauge-variant). So, while it would be ideal to work only with gauge-invariant metrics, we have very few metrics (and developing them and championing their adoption to the scientific community is beyond the scope of this work). Instead, to generate meaningful metrics, we gauge optimize gate sets to make them as “close” as possible before computing metrics.

Given a gate set and a target , we transform by (as above) where is chosen to optimize some criterion of “closeness” between and . This is “gauge optimization”. The final output of GST is thus the gate set that is most similar to the target, according to some gauge-variant quantity, among a class of gauge-equivalent gate sets. In the work reported here, we minimize (for convenience) a weighted Frobenius distance:


where denotes the Frobenius norm, ranges over all gates in the set, and and are weighting factors. The weight ratio allows us to fine-tune the relative contributions of discrepancies in logic gates and in SPAM. This is important because their respective natural uncertainties are usually quite different; gates can be probed far more accurately than SPAM. Thus, typically, ; we weight the gate matrix elements more highly because they are known more precisely. We use an iterative numerical method to find an that minimizes this quantity.

Each gate set is gauge-optimized as a whole; we report all metrics using gates in a single gauge. It would be incorrect to separately optimize the gauge for different reported quantities (e.g. gauge-optimizing for the fidelity of a single gate and reporting each such best-fidelity separately). Finally, we note that the process of gauge optimization against a reference gate set is sufficient to solve gauge ambiguity issues. That is, any quantity of interest that is not inherently gauge invariant (e.g., diamond norm) becomes so when this gauge optimization is performed. This numerical optimization process is not physically elegant, but is adequate for the practical applications we consider here.

iv.6 Error bars

In interpreting the GST analyses (and in particular confirming the claim that we have demonstrated fault tolerance) it is necessary to assign error bars to gate set estimates (and derived quantities thereof). For most GST-derived quantities, we use Hessian-based likelihood ratio (LR) confidence regions, while for RB-related quantities, we use parametric and non-parametric bootstrapping. We also use parametric bootstrapped error bars as a sanity check on our Hessian-based LR confidence regions, and find them to be in good agreement. Unless otherwise stated, all error bars indicate () confidence intervals.

We employ two flavors of bootstrapping: parametric and non-parametric. Both derive statistical quantities of interest from ensembles of simulated data sets, but these data sets are generated in different ways.

For the parametric bootstrap, ensembles of data sets are generated by first computing the GST estimate of the experimental data set in question, and then using this estimate to generate an ensemble of new data sets, each of which has the same experiments and number of shots per experiment as the actual experimental data set.

For the non-parametric bootstrap, ensembles of data sets are generated by simply resampling the experimental data set with replacement. In both parametric and non-parametric bootstraps, we typically generate an ensemble of 100 data sets, to ensure good statistics.

GST is used to map each resampled dataset to a gate set estimate. Each gate set is gauge-optimized to match the experimental GST estimate as closely as possible. Then, from this ensemble of gauge-optimized gate sets, any statistical quantity of interest (such as standard deviation) may be calculated for process matrix elements or for derived quantities such as diamond norm.

We use bootstrapped error bars for two purposes. First, they serve as a sanity check on the more rigorous (but tricky to implement) LR confidence regions described in the following subsection. In Figure 8 we compare the performance of parametric bootstrapping to the LR method, and see good agreement. Second, we use bootstrapping to put error bars on experimental RB decay rates. These are model-free and therefore not amenable to LR confidence regions. Error bars on experimental RB decay rate error bars were calculated via non-parametric bootstrapping, while error bars on simulated RB decay rates were calculated via parametric bootstrapping on the underlying GST estimate used to generate the RB data.

Figure 8: Comparison of error bar generation techniques. When computing error bars on GST estimates in this manuscript, we typically use likelihood ratio (LR) confidence regions, computed using the Hessian of the log likelihood function. However, another common approach is through parametric bootstrapping. Here, we show a log-log scatter plot of error bars on gate elements from GST estimate of our data from 30 March 2015. The x-axis corresponds to error bars calculated via parametric bootstrapping, whereas the y-axis corresponds to likelihood ratio (LR) confidence regions computed using the Hessian of the log likelihood function. The dotted line corresponds to . Both methods are described in the methods section “Error Bars”. The strong correlation shown here demonstrates the consistency between parametric bootstrapping techniques and LR confidence regions.

Bootstrapping is a very general method for generating error bars, but it is (1) not always reliable, (2) subject to small-sample errors unless very many Monte Carlo samples are generated, and (3) quite time-consuming (up to 24 hours of computer time were required to generate the 100 samples used for this paper). Likelihood ratio (LR) confidence regions RBK12 () are preferable in most ways, and we use them as our primary source for “error bars”.

The basic theory for LR confidence regions, as applied to quantum tomography, can be found in Ref. RBK12 (). Confidence regions have a solid (if often misunderstood) statistical meaning: if an estimator generates confidence regions with a confidence level of , then with probability at least (taken over the ensemble of all possible datasets), the confidence region assigned by the estimator will contain the true parameter value. This does not mean “Given particular error bars, the probability that they contain the truth is ,” as there is no random variable to take a probability with respect to once the estimate has been assigned.

As implemented here, GST has two convenient properties. First, it yields a likelihood function that is well-approximated by a Gaussian (because the total number of samples is quite large). Second, it involves no explicit constraints, meaning that the MLE is never squashed against a boundary (as it often is in standard state and process tomography, where the positivity constraint is critical to ensuring a physically valid estimate). These properties mean that we can approximate the loglikelihood function by a quadratic function, whose shape is given by the Hessian (matrix of 2nd derivatives) of at the MLE. This Hessian defines a covariance tensor in gate set space, which (when scaled by an appropriate factor) defines an ellipsoid that is a valid confidence region.

Writing down this ellipsoid explicitly (as a covariance tensor) is possible, but not useful in practice. Instead, we use it to define error bars (confidence intervals) for all relevant scalar quantities (including fidelities, diamond norms, gate matrix elements, etc).

Let be a scalar function of a gate set. We define a 95% confidence interval around the best-estimate value of by computing


where is the Hessian projected onto the (local) space of non-gauge gate set parameters, and we have linearized . is a scalar constant which satisfies , where is the cumulative density function of the probability distribution. With so defined, specifies the 95% confidence interval for . Within the linear approximation to , which is valid for small , this interval corresponds to minimizing and maximizing the value of over the contour of the log-likelihood corresponding to a 95% confidence interval if the log-likelihood had a single parameter.

We emphasize that this does not construct a 95% confidence region. There are roughly 34 gauge-invariant parameters in a gate set; the threshold used here implies 95% confidence intervals for each of them. The resulting region contains the truth only if every one of the intervals contains its parameter, which occurs with probability at least .

We believe this is a more meaningful way to report “error bars” than to report a 95% confidence region for the entire gate set. For one thing, it is consistent with the error bars reported by the bootstrap (which yields standard errors for each parameter independently, and would have to be expanded significantly to represent a joint confidence region). Empirically, we find that definition 27 correlates closely with the error bars on gate elements computed by parametric bootstrapping (see Fig. 8). Furthermore, we use the confidence region primarily to report uncertainties on single quantities (e.g. diamond norms), independent of the others.

iv.7 Data availability

The GST and RB analysis in this paper was performed using the open-source software pyGSTi (python Gate Set Tomography implementation) pygsti (), which was developed for this work. All datasets and analysis scripts necessary to reproduce the results presented here are available online as supplemental information at


  • (1) Nakamura, Y., Pashkin, Y. A. & Tsai, J. Coherent control of macroscopic quantum states in a single-cooper-pair box. Nature 398, 786–788 (1999).
  • (2) Wallraff, A. et al. Strong coupling of a single photon to a superconducting qubit using circuit quantum electrodynamics. Nature 431, 162–167 (2004).
  • (3) Barends, R. et al. Superconducting quantum circuits at the surface code threshold for fault tolerance. Nature 508, 500–3 (2014).
  • (4) Petta, J. R. et al. Coherent manipulation of coupled electron spins in semiconductor quantum dots. Science 309, 2180–4 (2005).
  • (5) Shulman, M. D. et al. Demonstration of entanglement of electrostatically coupled singlet-triplet qubits. Science 336, 202–205 (2012).
  • (6) Veldhorst, M. et al. A two-qubit logic gate in silicon. Nature 526, – (2015).
  • (7) Leibfried, D. et al. Experimental demonstration of a robust, high-fidelity geometric two ion-qubit phase gate. Nature 422, 412–415 (2003).
  • (8) Barrett, M. et al. Deterministic quantum teleportation of atomic qubits. Nature 429, 737–739 (2004).
  • (9) Reiserer, A., Kalb, N., Rempe, G. & Ritter, S. A quantum gate between a flying optical photon and a single trapped atom. Nature 508, 237–240 (2014).
  • (10) O’Brien, J. L., Pryde, G. J., White, A. G., Ralph, T. C. & Branning, D. Demonstration of an all-optical quantum controlled-not gate. Nature 426, 264–267 (2003).
  • (11) Peruzzo, A. et al. Quantum walks of correlated photons. Science 329, 1500–1503 (2010).
  • (12) Politi, A., Matthews, J. C. & O’Brien, J. L. Shor’s quantum factoring algorithm on a photonic chip. Science 325, 1221–1221 (2009).
  • (13) Monz, T. et al. 14-qubit entanglement: Creation and coherence. Phys. Rev. Lett. 106, 130506 (2011).
  • (14) Debnath, S. et al. Demonstration of a small programmable quantum computer with atomic qubits. Nature 536, 63–66 (2016).
  • (15) Harty, T. P. et al. High-fidelity preparation, gates, memory, and readout of a trapped-ion quantum bit. Phys. Rev. Lett. 113, 220501 (2014).
  • (16) Brown, K. R. et al. Single-qubit-gate error below in a trapped ion. Phys. Rev. A 84, 030303 (2011).
  • (17) Mount, E. et al. Error compensation of single-qubit gates in a surface-electrode ion trap using composite pulses. Phys. Rev. A 92, 060301 (2015).
  • (18) Lidar, D. & Brunn, T. (eds.) Quantum Error Correction (Cambridge University Press, 2013).
  • (19) Knill, E. Quantum computing with realistically noisy devices. Nature 434, 39–44 (2005).
  • (20) Aliferis, P., Gottesman, D. & Preskill, J. Quantum accuracy threshold for concatenated distance-3 codes. Quantum Info. Comput. 6, 97–165 (2006).
  • (21) Aliferis, P. & Cross, A. W. Subsystem fault tolerance with the bacon-shor code. Phys. Rev. Lett. 98, 220502 (2007).
  • (22) Aliferis, P. & Preskill, J. Fibonacci scheme for fault-tolerant quantum computation. Phys. Rev. A 79, 012332 (2009).
  • (23) Aliferis, P. et al. Fault-tolerant computing with biased-noise superconducting qubits: a case study. New J. Phys. 11, 013061 (2009).
  • (24) Aharonov, D. & Ben-Or, M. Fault-tolerant quantum computation with constant error rate. SIAM J. Comput. 38, 1207–1282 (2008).
  • (25) Emerson, J. et al. Symmetrized characterization of noisy quantum processes. Science 317, 1893–1896 (2007).
  • (26) Knill, E. et al. Randomized benchmarking of quantum gates. Phys. Rev. A 77, 012307 (2008).
  • (27) Sanders, Y. R., Wallman, J. J. & Sanders, B. C. Bounding quantum gate error rate based on reported average fidelity. New J. Phys. 18, 012002 (2016).
  • (28) Kueng, R., Long, D. M., Doherty, A. C. & Flammia, S. T. Comparing experiments to the fault-tolerance threshold. Phys. Rev. Lett. 117, 170502 (2016).
  • (29) Magesan, E. et al. Efficient measurement of quantum gate error by interleaved randomized benchmarking. Phys. Rev. Lett. 109, 080505 (2012).
  • (30) Kimmel, S., da Silva, M. P., Ryan, C. A., Johnson, B. R. & Ohki, T. Robust extraction of tomographic information via randomized benchmarking. Phys. Rev. X 4, 011050 (2014).
  • (31) Wallman, J., Granade, C., Harper, R. & Flammia, S. T. Estimating the coherence of noise. New J. Phys. 17, 113020 (2015).
  • (32) Blume-Kohout, R. et al. Robust, self-consistent, closed-form tomography of quantum logic gates on a trapped ion qubit. arXiv:1310.4492 (2013).
  • (33) Greenbaum, D. Introduction to quantum gate set tomography. arXiv:1509.02921 (2015).
  • (34) Kim, D. et al. Microwave-driven coherent operation of a semiconductor quantum dot charge qubit. Nat. Nanotechnol. 10, 243–247 (2015).
  • (35) Chuang, I. L. & Nielsen, M. Prescription for experimental determination of the dynamics of a quantum black box. J. Mod. Opt. 44, 2455–2467 (1997).
  • (36) Weinstein, Y. S. et al. Quantum process tomography of the quantum fourier transform. The Journal of chemical physics 121, 6117–6133 (2004).
  • (37) O’Brien, J. L. et al. Quantum process tomography of a controlled-not gate. Phys. Rev. Lett. 93, 080502 (2004).
  • (38) Shulman, M. D. et al. Demonstration of entanglement of electrostatically coupled singlet-triplet qubits. Science 336, 202–205 (2012).
  • (39) Merkel, S. T. et al. Self-consistent quantum process tomography. Phys. Rev. A 87, 062119 (2013).
  • (40) Olmschenk, S. et al. Manipulation and detection of a trapped hyperfine qubit. Phys. Rev. A 76, 052314 (2007).
  • (41) Wimperis, S. Broadband, Narrowband, and Passband Composite Pulses for Use in Advanced NMR Experiments. J. Magn. Reson., Ser A 109, 221–231 (1994).
  • (42) Merrill, J. T. & Brown, K. R. Progress in Compensating Pulse Sequences for Quantum Computation, 241–294 (John Wiley & Sons, Inc., 2014).
  • (43) Gustavsson, S. et al. Improving quantum gate fidelities by using a qubit to measure microwave pulse distortions. Phys. Rev. Lett. 110, 040502 (2013). URL
  • (44) Morton, J. J. L. et al. Measuring errors in single-qubit rotations by pulsed electron paramagnetic resonance. Phys. Rev. A 71, 012332 (2005).
  • (45) O’Malley, P. J. J. et al. Qubit metrology of ultralow phase noise using randomized benchmarking. Phys. Rev. Applied 3, 044009 (2015).
  • (46) Kelly, J. et al. Optimal quantum control using randomized benchmarking. Phys. Rev. Lett. 112, 240504 (2014).
  • (47) Stick, D. et al. Demonstration of a microfabricated surface electrode ion trap. arXiv:1008.0990 (2010).
  • (48) Magesan, E., Blume-Kohout, R. & Emerson, J. Gate fidelity fluctuations and quantum process invariants. Phys. Rev. A 84, 012309 (2011).
  • (49) Magesan, E., Gambetta, J. M. & Emerson, J. Characterizing quantum gates via randomized benchmarking. Phys. Rev. A 85, 042311 (2012).
  • (50) Khodjasteh, K. & Viola, L. Dynamical quantum error correction of unitary operations with bounded controls. Phys. Rev. A 80, 032314 (2009).
  • (51) Aharonov, D., Kitaev, A. & Nisan, N. Quantum circuits with mixed states. In Proceedings of the Thirtieth Annual ACM Symposium on Theory of Computing, STOC ’98, 20–30 (ACM, New York, NY, USA, 1998).
  • (52) Benenti, G. & Strini, G. Computing the distance between quantum channels: usefulness of the fano representation. J. Phys. B: At., Mol. Opt. Phys. 43, 215508 (2010).
  • (53) Watrous, J. Simpler semidefinite programs for completely bounded norms. Chicago Journal of Theoretical Computer Science 8, 11–14 (2013).
  • (54) Wasserman, L. All of Statistics: A Concise Course in Statistical Inference. Springer Texts in Statistics (Springer New York, 2013).
  • (55) Wallman, J. J. & Flammia, S. T. Randomized benchmarking with confidence. New J. Phys. 16, 103032 (2014).
  • (56) Ball, H., Stace, T. M., Flammia, S. T. & Biercuk, M. J. Effect of noise correlations on randomized benchmarking. Phys. Rev. A 93, 022303 (2016).
  • (57) Wallman, J. J. Bounding experimental quantum error rates relative to fault-tolerant thresholds. arXiv:1511.00727 (2015).
  • (58) Wallman, J. J. & Emerson, J. Noise tailoring for scalable quantum computation via randomized compiling. Phys. Rev. A 94, 052325 (2016).
  • (59) Jones, E., Oliphant, T., Peterson, P. et al. SciPy: Open source scientific tools for Python (2001–). URL
  • (60) Blume-Kohout, R. Robust error bars for quantum tomography. arXiv:1202.5270 (2012).
  • (61) Nielsen, E., Rudinger, K., Gamble, J. K. & Blume-Kohout, R. pyGSTi: A python implementation of gate set tomography (2016). URL

Appendix A Acknowledgements

Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-AC04-94AL85000. The authors thank Travis Scholten and Jonathan Gross for assistance with data visualization and Kevin Young for providing code for diamond norm computation. JKG gratefully acknowledges support from the Sandia National Laboratories Truman Fellowship Program, which is funded by the Laboratory Directed Research and Development (LDRD) program. This research was funded, in part, by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA). All statements of fact, opinion or conclusions contained herein are those of the authors and should not be construed as representing the official views or policies of IARPA, the ODNI, or the U.S. Government.

Appendix B Author contributions

RBK, JKG, EN, and KR contributed to the theoretical development of GST and the code implementation. JM, KF, and PM performed the experiments. All authors discussed the results and wrote the manuscript.

Appendix C Competing interests

The authors declare no competing financial interests.

Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description