This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.
Paper The following article is Open access

Handling leakage with subsystem codes

, and

Published 30 July 2019 © 2019 The Author(s). Published by IOP Publishing Ltd on behalf of the Institute of Physics and Deutsche Physikalische Gesellschaft
, , Citation Natalie C Brown et al 2019 New J. Phys. 21 073055 DOI 10.1088/1367-2630/ab3372

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

1367-2630/21/7/073055

Abstract

Leakage is a particularly damaging error that occurs when a qubit state falls out of its two-level computational subspace. Compared to independent depolarizing noise, leaked qubits may produce many more configurations of harmful correlated errors during error-correction. In this work, we investigate different local codes in the low-error regime of a leakage gate error model. When restricting to bare-ancilla extraction, we observe that subsystem codes are good candidates for handling leakage, as their locality can limit damaging correlated errors. As a case study, we compare subspace surface codes to the subsystem surface codes introduced by Bravyi et al. In contrast to depolarizing noise, subsystem surface codes outperform same-distance subspace surface codes below error rates as high as ⪅ 7.5 × 10−4 while offering better per-qubit distance protection. Furthermore, we show that at low to intermediate distances, Bacon–Shor codes offer better per-qubit error protection against leakage in an ion-trap motivated error model below error rates as high as ⪅ 1.2 × 10−3. For restricted leakage models, this advantage can be extended to higher distances by relaxing to unverified two-qubit cat state extraction in the surface code. These results highlight an intrinsic benefit of subsystem code locality to error-corrective performance.

Export citation and abstract BibTeX RIS

1. Introduction

Quantum information is subject to many different noise processes that depend upon the platform in which it is physically encoded. Quantum error-correcting codes can be used to preserve such information for any length computation, given certain assumptions on the strength and correlations of the error processes [1, 2]. In this work, we address an error model pertinent to many qubit architectures, leakage, and identify subsystem codes well-suited to correct it.

1.1. Leakage

Leakage is a particularly damaging error that occurs when a qubit excites to a state outside of its two-level computational subspace [310]. Much work has been devoted to characterizing and reducing leakage at the experimental level in different qubit architectures, including photonics [11], quantum dots [1217], superconductors [1828], topological qubits [29], and ion traps [3032]. Unlike erasure, the locations of leakage errors are unknown, and so it is far more damaging [3335].

Leakage rates are significant for several types of qubits [31, 36], and can even be the dominant error rate in the system [11, 37]. However, even when leakage is relatively less likely than other types of noise, the havoc it wreaks at the level of error-correction can quickly make it a limiting error process [31, 38]. This is for two reasons. First, standard error-correction is not equipped to handle errors acting outside the computational subspace, and so leakage can persist until it is corrected by some other means. Second, before it is corrected, leakage can cause damaging patterns of correlated errors that depend on the specifics of a platform's gate implementations.

Leakage can also present a fundamental error when compared to other noise sources. For example, while improved controls are continually reducing the dominating effects of coherent errors in ion-trap technologies, scattering events that cause leakage are more difficult to subdue. Consequently, although lower leakage rates make it a lesser concern for certain architectures currently, handling leakage in the low-error regime is vitally important for performing long computations reliably in the future.

Fortunately, one can reduce leakage by regularly swapping or teleporting qubits to the computational subspace [38, 39], and it has been rigorously shown that a threshold exists in this model [40]. This introduces a balancing act. Eliminating leakage too frequently incurs significant qubit and gate overhead, which in turn introduces more potential circuit faults. Eliminating leakage too infrequently allows leaked qubits to interact with many other qubits, introducing higher-weight correlated errors.

1.2. Local codes

In this work, we focus on leakage in geometrically local codes, which are leading candidates for implementing quantum memories. Included in this category are topological codes, which exhibit some of the highest thresholds, support minimal overhead syndrome extraction, and are naturally amenable to qubit architectures that prefer local interactions [41].

Local codes are also a natural candidate for coping with leakage, as the number of qubits that may interact with any one leakage event is limited. Consequently, previous works have focused on quantifying the performance of topological subspace codes in the presence of leakage, including the surface code [13, 23, 31, 38, 42] and [[7, 1, 3]] Steane or color code [11].

The (subspace) surface code is a particularly enticing choice as, in many respects, it is the top-performing quantum memory and a quantum analog of the repetition code [43, 44]. This analogy can be made precise, in the sense that the surface code nearly saturates the CSS zero-rate entropic bound $2H({p}_{\mathrm{thr}})\leqslant 1$ and can be realized as the hypergraph product of the repetition code with itself [4548]. Among its desirable features is the use of a single ancilla to extract syndromes, relevant to both its high performance and square lattice locality [41, 49].

However, the surface code still suffers from significantly reduced thresholds [38] and an effective distance reduction [31, 42] in the presence of leakage. It is then natural to ask if other local code families may perform better in certain regimes of a leakage error model.

1.3. Subsystem codes

Diverging from previous works, our focus will be on subsystem codes, and in particular, subsystem surface and Bacon–Shor codes. Subsystem codes are a generalization of subspace codes in which certain logical degrees of freedom are treated as gauges, and allowed to vary freely [50, 51]. Unlike topological subspace stabilizer codes, which require at least 4-local interactions in a nearly Euclidean lattice [52], topological subsystem codes can be realized with 3- or even 2-local interactions [5356]. If one sacrifices topological protection, then non-topological local codes like the Bacon–Shor family can still yield good error protection at reasonable error rates with only 2-local interactions and minimal overhead [57, 58].

Generally, subsystem codes come with certain intrinsic advantages. Their increased locality is even more desirable for physically local qubit architectures. They allow for simpler and parallelized syndrome extraction, which mitigates the circuit-depth required to measure higher-weight stabilizers [59]. Geometrically constrained codes can yield better encoding rates compared to their subspace cousins [44, 48, 60]. Finally, subsystem codes can yield tremendous advantages for implementing simple and universal fault-tolerant logic [6163], bypassing the high cost of magic-state distillation [64].

However, the advantages of subsystem code locality often come at a high cost to their error-corrective performance. While gauge degrees of freedom are useful for implementation in local architectures and fault-tolerant computation, they introduce more locations for potential circuit faults without giving additional information about the errors that are occurring. This generally manifests as less reliable high-weight stabilizers and lower thresholds.

1.4. Contributions

In this work, we highlight a simple and intrinsic error-corrective advantage for these gauge degrees of freedom: limiting correlated errors due to leakage. This differs from an independent depolarizing model, in which local measurements do not reduce the set of correlated errors.

To be explicit, suppose we wanted to measure a stabilizer $S={G}_{1}\ldots {G}_{{\ell }}$, where each Gi is a gauge operator. If we measure each gauge separately, then a single fault while measuring Gi will produce an error supported entirely on the support of Gi. If instead we measure the entire stabilizer directly, then a single fault could produce an error supported on the support of many Gi. However, in an independent depolarizing model, as long as we measure along the gauge operators, this error will take the form ${{EG}}_{i+1}\ldots {G}_{n}$ where E is supported entirely on the support of Gi. In particular, the set of correlated errors is the same with or without local measurements up to a symmetry of the logical subsystem. For some gate-error models with non-local interactions, it is even advantageous to measure the larger stabilizers directly [46, 65].

However, this is not true for typical leakage models, as a single leakage can produce a much richer set of correlated errors. Consequently, circuits that are fault-tolerant in a depolarizing model may no longer be fault-tolerant in a leakage model. We illustrate this difference in the single ancilla extraction of a Bacon–Shor code stabilizer with both Pauli and leakage faults in figure 1.

Figure 1.

Figure 1. Syndrome extraction in a vertical distance 3 X-type Bacon–Shor stabilizer (shaded) using a single ancilla [65], with horizontal XX and vertical ZZ gauge operators. Implicit in the diagram is a Pauli X-error on the ancilla in (a), and a leaked ancilla in (b), both occurring after the third gate during extraction. The correlated error in (a) is equivalent to a single data qubit X-error up to gauge transformation. If, for example, the leaked ancilla in (b) depolarizes each qubit it interacts with, then it may produce a damaging higher-weight correlated error like the one shown.

Standard image High-resolution image

We compare three code families in the presence of leakage: subspace surface codes, subsystem surface codes, and Bacon–Shor codes, while restricting to bare-ancilla syndrome extraction. We consider the first two in both standard and rotated lattice geometries, and under various leakage reduction techniques. In stark contrast to depolarizing noise, we find that subsystem surface codes actually outperform their same-distance subspace cousins at equal leakage-to-depolarizing ratios in the low-error regime. Furthermore, we observe that low to intermediate distance Bacon–Shor codes outperform subspace surface codes in per-qubit leakage protection at similar error rates. These advantages are due essentially to the subsystem codes' handling of uncontrolled 'hook' errors during syndrome extraction with leaked qubits. Finally, if one allows for larger ancilla states in a restricted leakage model, we observe that intermingling minimal leakage reduction with partial fault-tolerance in the subspace code ultimately yields the best overall low-error performance. Our central contributions are the following.

  • (i)  
    We motivate a Pauli-approximated ion-trap leakage model, and demonstrate significantly improved performance compared to a worst-case stochastic leakage model. This model is similar to a photonics leakage model [11].
  • (ii)  
    We show that any topological subspace stabilizer code must incur a linear reduction in effective distance when using minimal leakage reduction to correct worst-case stochastic leakage. In the case of subspace surface codes, the effective code distance is halved [31, 38, 42].
  • (iii)  
    We define a rotated variant of the subsystem surface codes proposed in [53] which yields reduced qubit overhead. We argue that these codes (along with their un-rotated variant) yield better per-qubit distance protection than subspace surface codes in the presence of leakage. We give coarse upper-bounds on the error regime in which this advantage manifests, with $p\lessapprox 2.5\times {10}^{-4}$ with worst-case stochastic leakage and $p\lessapprox 0.32\times {10}^{-4}$ with ion-trap leakage.
  • (iv)  
    We provide Monte Carlo simulations of both the subspace and subsystem surface codes at low error rates using no qubit overhead for leakage elimination. We find that subsystem surface codes outperform same-distance, same-geometry subspace surface codes for error rates $\lessapprox 0.75\times {10}^{-3}$ with worst-case stochastic leakage. This is increased to $\lessapprox 2.0\times {10}^{-3}$ for ion-trap leakage.
  • (v)  
    We additionally compare Bacon–Shor codes with serialized syndrome extraction to subspace surface codes with ion-trap leakage. We observe that at low distances 3, 5, and 7, Bacon–Shor codes per-qubit outperform surface codes in the ${10}^{-4}\mbox{--}{10}^{-3}$ error range. This advantage persists even when the depolarizing rate is many times stronger than the leakage rate.
  • (vi)  
    Finally, we compare different leakage reduction techniques in an ion-trap leakage model of the surface code while allowing for larger ancilla states. We observe that combining unverified two-qubit cat state extraction with minimal leakage reduction yields an orders-of-magnitude improvement over existing fault-tolerant proposals [38], and also begins to relatively outperform non-fault-tolerant strategies in the $\lessapprox {10}^{-3}$ error range. This advantage may extend to higher distances, but relies unavoidably on an assumption of independent leakage events.

At the heart of these results is a simple accounting of the correlated errors that may appear due to leakage of both data and ancilla qubits, and the quantification of the regime in which their effects dominate. However, it underscores an advantage of local subsystem error-correction that may extend to other non-Markovian error models, with local baths that are determined by the qubit connectivity [66]. In these cases, local checks minimize what can go wrong.

The paper proceeds as follows. In section 2, we define the two leakage models we consider, and introduce the leakage reduction techniques we will apply. In section 3, we compare the leakage robustness of subspace surface codes and subsystem surface codes by accounting for fault-paths generated by leakage. Although we focus on low-overhead leakage reduction with bare-ancilla measurements, we touch on more general leakage reduction strategies in appendix C. In section 4 we provide numerics comparing subspace and subsystem codes under various leakage reduction schemes and constraints. Finally, we conclude with some discussion and potential avenues for improvement in section 5. Additional technical details concerning the simulations and gate schedulings may also be found in appendix B.

2. Leakage

Qubits are not completely isolated two-level systems, and sometimes they have energetic access to other states. Leakage occurs when a qubit state excites out of its computational subspace.

We call a qubit leaked if its state is supported on the subspace ${{ \mathcal H }}_{| 2\rangle }$, where $| 2\rangle $ represents the leaked degree of freedom. Otherwise we call the qubit sealed. Note that there may be many leakage states, and leakage need not take this restricted form, but for simplicity we group these into a single $| 2\rangle $-state.

Although leakage may be modeled as a coherent process [21], we make a standard simplifying assumption [38] that leakage occurs stochastically via the channel

We refer to p as the leakage rate of the system. This simplified error channel ensures that we only consider mixtures of leakage states and sealed states, which will be necessary for efficient simulation. Additionally, we assume that each qubit undergoes a relaxation channel whenever it could also leak

where ${P}_{C}=| 0\rangle \langle 0| +| 1\rangle \langle 1| $ is the projector onto the computational space, and pr is the relaxation rate. For simplicity, we assume ${p}_{r}={p}_{l}$.

In addition to leakage, we also assume background depolarizing noise. In order to probe the leakage behavior of the codes, we do not want the depolarizing noise to drown out leakage effects. However, not including depolarizing noise introduces peculiarities in the error model, such as perfect measurement. Consequently, we choose our depolarizing rate pd to be the same as our leakage rate p, unless otherwise specified. An equal leakage-to-depolarizing ratio is directly applicable in certain systems [11, 31, 37], but even with depolarizing noise that is many times stronger, leakage may remain a prohibitive error source [31, 38]. We define the error rate of the system to be $p={p}_{{\ell }}+{p}_{d}$, as the leakage and depolarizing error processes occur independently.

Our gate error model involves three fundamental operations: preparation into $| 0\rangle $ or $| +\rangle $, a two-qubit CNOT gate, and measurement in the Z- and X-bases. It is important to note that our gate error model does not include idling errors, a choice we make to simplify gate scheduling. While idling errors do not usually cause leakage, one must be careful to consider more careful parallelization in architectures with shorter coherence times.

Preparation can fail by wrongly preparing an orthogonal state with probability pd. Additionally, a newly-prepared qubit can immediately leak with probability p; for some systems, this is a pessimistic assumption.

A two-qubit CNOT gate can fail by applying one of the 15 non-identical two-qubit Pauli operators to its support, each with probability ${p}_{d}/15$. For our simulations, we assume that leakage events occur with probability p independently on the support of the gate. This is well-motivated in some architectures [38], but less so in others. However, the subsystem code constructions can also handle correlated leakage events with small modification, which we discuss in appendix F.

Finally, measurements can return the wrong result with probability pd. Whenever a leaked qubit is measured, it will always return eigenvalue −1. Implicitly, this assumes two-level measurements that cannot distinguish between excited states, and is also relevant to many qubit architectures [26, 38]. However, while this error-model is used in simulating logical error rates, we consider an even more damaging leakage model when accounting for fault-paths to argue about effective distance. In this second model, measurements on leaked qubits produce randomized outcomes, matching [42].

2.1. Leakage models

There is some freedom in extending the dynamics of the system to interactions between leaked qubits and sealed qubits. In particular, we assume that our two-qubit gates are themselves sealed. For ${{ \mathcal H }}_{C}$ the computational Hilbert space and ${{ \mathcal H }}_{| 2\rangle }$ the leakage Hilbert space, this means that any two-qubit gate U factors as

This restriction is physically motivated in many architectures, including both superconductors [38, 42] and ion-traps [31]. In particular, U does not directly propagate one leaked qubit into two leaked qubits. Note that we can similarly define a sealed single-qubit gate as one that decomposes as $U={U}_{{{ \mathcal H }}_{C}}\oplus {U}_{{{ \mathcal H }}_{| 2\rangle }}$.

The two-qubit gate ${U}_{{{ \mathcal H }}_{C}\otimes {{ \mathcal H }}_{C}}$ is the computational gate and ${U}_{{{ \mathcal H }}_{| 2\rangle }\otimes {{ \mathcal H }}_{| 2\rangle }}$ can only apply a harmless phase. However, for two-qubit gates, it remains to define ${U}_{{{ \mathcal H }}_{C}\otimes {{ \mathcal H }}_{| 2\rangle }}$ and ${U}_{{{ \mathcal H }}_{| 2\rangle }\otimes {{ \mathcal H }}_{C}}$, the interactions between leaked and sealed qubits.

2.1.1. Depolarizing leakage

We consider two such definitions, resulting in two different leakage models. The first is a worst-case stochastic model that has been considered in previous works [38, 42]. We call this model depolarizing or DP-leakage, and it is defined symmetrically for both ${U}_{{{ \mathcal H }}_{C}\otimes {{ \mathcal H }}_{| 2\rangle }}$ and ${U}_{{{ \mathcal H }}_{| 2\rangle }\otimes {{ \mathcal H }}_{C}}$. In the DP-leakage model, the leaked qubit remains leaked while a Haar random unitary is applied to the sealed qubit. This results in complete depolarization of the sealed qubit, and is the more damaging of the two leakage models.

2.1.2. Mølmer–Sørensen leakage

The second is a more restrictive model in which a leaked qubit simply does not interact with a sealed qubit. This presents an effective gate-erasure error, which we expect may apply to several different architectures. One such example is the standard entangling operation for ion-traps: the Mølmer–Sørensen gate [67].

The two-qubit Mølmer–Sørensen interaction is mediated by the shared motional modes in an ion-trap. The underlying Hamiltonian is a state-dependent displacement, which is based on near-resonant driving of motional sidebands that correspond to removing or adding phonons of motion with excitation of the internal ion. The Hamiltonian parameters are chosen to remove any residual entanglement between the ion and motion at the end of the gate. The ion–ion entanglement arises from a second-order term due to the time-dependent displacement not commuting with itself in time.

Mølmer–Sørensen gates are often optimized in a way that is independent of the magnitude of the driving field, with the residual entanglement minimized ion by ion, and do not require a second ion to be driven. For a leaked ion, if the spacing between the motional modes is small compared to the separation between the computational energy state and the leakage energy state, then the lasers that drive the displacement are far from resonant and both to the same side of the carrier transition. The result is that the leaked ion is only very weakly displaced and therefore does not generate an entangling gate with the sealed ion, similar to a photonics model [11].

The Mølmer–Sørensen gate does not occur and the sealed ion is unaffected. For the full CNOT gate, the net effect is that different single qubit Pauli rotations are applied to the control and target qubits, see figure 2. The target undergoes an $X(-\pi /2)$-rotation along the Bloch sphere, while the control undergoes an $Z(-\pi /2)$-rotation. Whichever qubit is leaked is unaffected by the sealed single-qubit gate. In order to simulate leakage in the Pauli frame model in the same framework as [38, 42], we replace these channels with their Pauli-twirl approximations. This yields the stochastic error channels

where ${{ \mathcal E }}_{{\rm{bit}}}$ is applied to the target when the control is leaked, and ${{ \mathcal E }}_{{\rm{phase}}}$ is applied to the control when the target is leaked. We refer to this model as Mølmer–Sørensen or MS-leakage. The twirl of this restricted model will yield significantly improved performance compared to DP-leakage, although one must be careful to weigh this against the generic benefit of twirling [68, 69]3 .

Figure 2.

Figure 2. An ion-trap CNOT gate, expressed as a product of native Pauli rotations, where $P(\theta ):= \exp (-{\rm{i}}\theta P/2)$ for single-qubit Pauli operators P and ${XX}(\theta ):= $ $\exp (-{\rm{i}}\theta {XX})$. In the presence of a leaked qubit, the dashed Mølmer–Sørensen gate is not applied.

Standard image High-resolution image

2.2. Leakage reduction

As errors preserving the computational subspace accumulate, error-correction will periodically remove them. However, standard error-correction does nothing to remove leakage. Every leakage event will eventually relax back to the computational subspace, but these long-lived leakage errors will corrupt the surrounding qubits for several rounds of error-correction [42]. Without an active approach to remove leakage, it will completely compromise the efficacy of the code, see figure 3.

Figure 3.

Figure 3. Distance 3 (circle), 5 (cross), and 7 (triangle) standard surface code performance both with depolarizing leakage (dashed) at ${p}_{d}={p}_{r}=100{p}_{{\ell }}$ and without depolarizing leakage (solid) while using no leakage reduction. Spontaneous relaxation back to the computational subspace is an insufficient mechanism for leakage reduction even when leakage is orders of magnitude less likely than depolarization, and active methods are required.

Standard image High-resolution image

Leakage reduction units or LRUs are circuit gadgets that are used to actively remove leakage [39, 40]. They are defined by two properties:

  • (i)  
    if the quantum state is sealed, an LRU ideally acts as the identity on that state, and
  • (ii)  
    if the quantum state is leaked, then an LRU ideally projects the state back to the computational subspace.

There are two popular approaches for eliminating leakage. The first is to introduce auxiliary qubits and regularly swap between them and the initial data qubits in LRUs. After the states are swapped, the initial data qubits are measured and reprepared in the computational space, removing leakage. If one relaxes the assumption that the gates are sealed, then one must teleport the initial data qubits to remove leakage, see figure 4. The frequency and placement of these LRUs will determine the code performance, which we address in section 3.

Figure 4.

Figure 4. Leakage can be removed via LRUs constructed from swapping with an auxiliary qubit (top) or teleportation (bottom). In the ideal case, the data D is teleported to the bottom wire. Note that a single leakage cannot persist through the LRU in both the sealed-gate (top) and unsealed-gate (bottom) settings.

Standard image High-resolution image

The second approach foregoes auxiliary qubits, and instead periodically swaps the roles of data and ancilla in the code. This ensures that each physical qubit is measured in every other round [13, 31, 38]. This technique requires no qubit overhead, and may preserve the locality of the qubit lattice. For these reasons, it may be the more desirable approach. While it is more complicated to analyze the effects of longer-lived leakage faults, we study this technique numerically in section 4.

How frequently should we insert LRUs into our error-correction circuit? We focus on two minimal overhead strategies.

  • (i)  
    Syndrome extraction leakage reduction (syndrome-LR) swaps each data qubit with an associated auxiliary qubit at the end of each syndrome extraction. This allows a single leakage to persist for a single syndrome extraction cycle. With some circuit delay, the ancillae may be used as the auxiliary qubits.
  • (ii)  
    Swap leakage reduction (swap-LR) removes leakage by applying a swap gate between each ancilla qubit and the last data qubit it interacts with in each syndrome extraction cycle. This allows a single leakage to persist for at most two consecutive syndrome extraction cycles; see figure 10.

3. Leakage robustness

In this section, we detail damaging correlated errors due to leakage, and argue about the effective distance of different surface codes using syndrome-LR. We compare subspace and subsystem surface codes, in both standard and rotated lattice geometries, in the presence of both DP- and MS-leakage using bare-ancilla extraction. While we focus on leakage faults using syndrome-LR, we give numerical evidence that the additional time-correlated errors introduced by swap-LR are not too damaging.

In each case, we consider planar codes with boundary encoding a single qubit4 . Intuitively, the increased locality of the subsystem codes limits the correlated errors that can occur due to leakage. This comes at the expense of a larger qubit lattice to achieve the same code distance. We also allow lattice geometries that are either rotated (a periodic diamond cut) or standard (i.e. un-rotated, a periodic square cut). Rotated surface codes give a $\sqrt{2}$-fold increase in code distance [49, 70] per qubit. However, a higher fraction of fault-patterns may be more damaging, and recent work has shown that standard surface codes may outperform rotated surface codes within certain sub-threshold error regimes [71]. A similar intuition holds for leakage, as a standard lattice geometry can ensure that leakage faults do not propagate errors parallel to a logical operator, again at the cost of additional qubit overhead.

We call a code leakage robust if it does not experience a linear effective distance reduction in the presence of leakage. Otherwise, we call them leakage susceptible, in which case the effective distance is halved in surface codes. To achieve a desired ${d}_{\mathrm{eff}}$ in the presence of MS-leakage, we find that it is asymptotically optimal to choose a rotated subsystem code, and next a standard subspace code, both of distance ${d}_{\mathrm{eff}}$. In the presence of DP-leakage, we find that it is asymptotically optimal to choose a standard subsystem code of distance ${d}_{\mathrm{eff}}$, and next a rotated subspace code of distance $2{d}_{\mathrm{eff}}$.

In both cases, the subsystem surface codes yield better per-qubit distance protection than subspace surface codes, as summarized in table 1. We expect this subsystem advantage to generalize to multiple encoded qubits and persist against any surface code geometry, as rotated [[n, k, d]] surface codes already saturate the 2D topological code bound ${{kd}}^{2}\leqslant {cn}$, where $c\geqslant 1$ for a planar square lattice architecture with closed defects [44, 72].

Table 1.  Total number of data qubits required to realize an effective distance d in the presence of different leakage models. Accounting for ancilla qubits depends on tradeoffs between qubit reuse and parallelization. In the fully parallelized case, the subsystem advantage persists but is reduced; see appendix C.

  DP-Leakage MS-Leakage
Rotated subspace $4{d}^{2}+O(d)$ $4{d}^{2}+O(d)$
Rotated subsystem $6{d}^{2}+O(d)$ $1.5{d}^{2}+O(d)$
Standard subspace $8{d}^{2}+O(d)$ $2{d}^{2}+O(d)$
Standard subsystem $3{d}^{2}+O(d)$ $3{d}^{2}+O(d)$

Better per-qubit distance protection implies that subsystem surface codes outperform subspace surface codes in the $p\to 0$ limit. Although thresholds may also be reduced by the addition of local correlated data qubit errors, these entropic effects are relatively mild [73]. Consequently, at higher error rates, subspace surface codes achieve better performance due to their higher threshold even if the effective distance is reduced. After analyzing correlated fault patterns using syndrome-LR in the present section, we investigate this crossover point in same-distance codes using swap-LR in section 4 numerically.

It is worth noting that the insertion of additional leakage reduction can restore the effective distance of any code. For example, performing a leakage reduction step after each individual gate immediately converts leakage errors to depolarizing errors as they arise. However, this introduces enormous circuit-volume overhead, which results in significantly reduced performance and many extra qubits [38]. Additionally, unlike swap-LR, it will necessarily increase the required connectivity of the qubit lattice. Thus, we focus on minimal overhead leakage reduction, but discuss additional strategies in appendix C.

3.1. Subspace surface codes

We first consider fault-tolerance using syndrome-LR on subspace surface codes with bare-ancilla syndrome extraction. Subspace surface codes are defined on a square lattice, with qubits placed on the edges of the lattice. To each plaquette P in the bulk of the lattice, we associate a 4-body X-type stabilizer, and to each vertex V in the lattice, we associate a 4-body Z-type stabilizer,

Then the stabilizer group is generated by all plaquette and vertex operators of this form, as well as those on the boundary. For a standard lattice these boundary stabilizers are weight-3, while for rotated lattices they are weight-2, and we refer to violated stabilizers in the lattice as excitations. Standard codes form a family with parameters $[[2{d}^{2}-2d+1,1,d]]$, while rotated codes form a family with parameters $[[{d}^{2},1,d]]$.

In the presence of DP-leakage, both the standard and rotated surface codes experience a halving of their effective distance, caused in part by uncontrolled 'hook' errors [41] during syndrome extraction [31, 38, 42]. Three such distance-damaging faults, caused by ancilla leakage, are shown in figure 5.

Figure 5.

Figure 5. Distance-damaging leakage faults in the distance 3 standard and rotated surface codes. The red L indicates a DP-leakage event; the blue L indicates an MS-leakage event, and anticommuting excitations are filled by the same color. Here, within the random measurement outcome leakage model of [42], leaked qubits are measured as +1. The rotated surface code has north–south X-type boundaries, and east–west Z-type boundaries.

Standard image High-resolution image

However, standard subspace codes are robust to MS-leakage. Unlike depolarizing errors, a single data leakage event may cause many different configurations of measurement outcomes on incident stabilizer checks before its removal. Thus, a space-correlated error due to a data leakage may produce any one of its possible time-correlated syndrome configurations, but only over a single time-step. Given d successive syndrome measurements, this reduces the problem to considering only space-correlations in fault patterns.

In the MS-leakage model, data leakage is not too damaging: although it may generate many combinations of time-correlated syndrome configurations, it does not produce any new space-correlated errors. The reason is that data MS-leakage cannot propagate errors to ancillae that will then propagate errors to other data qubits. Thus, it suffices to only consider space-correlated errors due to ancilla leakage.

Ancilla MS-leakage may produce arbitrary configurations of errors on the support of a stabilizer that are of the same type as that stabilizer. For the subspace surface code, every such configuration is either a single-qubit or two-qubit error, up to stabilizer equivalence. In particular, in the standard lattice, there are two different configurations of excitations (up to symmetry) caused by weight-2 correlated errors. Both are contained in an ${L}^{\infty }$-ball of radius one on the ancilla qubit lattice, and so any configuration of excitations caused by $\lt d$ faults cannot traverse the lattice. As any logical operator which does not traverse the lattice must be trivial, we may conclude that the effective distance of the code is preserved. See figure 6 for a summary of these worst-case leakage events.

Figure 6.

Figure 6. Worst-case MS-leakage in subspace surface codes that does not cause an effective distance reduction, with Z-type excitations colored red. Again, we assume north–south X-type boundaries of the rotated code. Note that the diagonal errors in the standard lattice geometry may turn parallel to a logical operator in the rotated lattice geometry.

Standard image High-resolution image

3.2. Leakage susceptibility of subspace codes

As subspace surface codes are susceptible to DP-leakage, each experiences an effective distance reduction $d\mapsto \lceil \tfrac{d}{2}\rceil $ [31, 38, 42]. This is damaging to their low-error suppression, and we will show that this damage begins to manifest at low but relevant sub-threshold error rates. One might hope to construct some clever topological code and scheduling that minimizes the damage that DP-leakage inflicts on the effective code distance.

Unfortunately, a simple union bound shows that this damage is inevitable when restricting to minimal overhead leakage reduction in topological stabilizer subspace codes. However, the proof will give intuition for avoiding this damage by extending the search to topological stabilizer subsystem codes.

(Union Bound).

Proposition 1 Let $\{{C}_{d}\}$ be a family of D-dimensional topological subspace stabilizer codes parameterized by diverging distance d using bare-ancilla syndrome extraction and syndrome-LR. Let ${d}_{\mathrm{eff}}$ denote the minimum number of circuit faults that may cause a logical error in the presence of DP-leakage. Then there must be a linear effective distance reduction, i.e. there is a constant $\eta \lt 1$ such that ${d}_{\mathrm{eff}}\leqslant \eta d.$

Proof. By definition, each Cd is defined on a D-dimensional lattice of qubits with a metric. As the code family is topological, there must be some maximum diameter ξ of the support of any stabilizer generator of the code, where ξ is independent of d.

Select any weight-d logical operator

acting on qubits ${{\ell }}_{1},\,\ldots ,\,{{\ell }}_{d}$, where each $P\in \{X,Y,Z\}$. Select a maximal set R of disjoint neighborhoods ${B}_{{{\ell }}_{i}}(\xi )$ centered at qubits in the support of L. Recall that

Then there must be at least $\lfloor \tfrac{d}{{c}_{D}{\xi }^{D}}\rfloor $ such neighborhoods in this collection.

Using syndrome-LR, note that r ancilla leakage events can cause any configuration of errors on the combined support of r stabilizer generators. Furthermore, for each ${B}_{{{\ell }}_{i}}(\xi )\in R$, there is some stabilizer generator ${S}_{{{\ell }}_{i}}$ such that $\{{P}_{{{\ell }}_{i}},{S}_{{{\ell }}_{i}}\}=0$. As $[{S}_{{{\ell }}_{i}},L]=0$, there must be at least one other ${P}_{{{\ell }}_{j}}$ supported in ${B}_{{{\ell }}_{i}}(\xi )$ with $\{{P}_{{{\ell }}_{j}},{S}_{{{\ell }}_{i}}\}=0$.

In particular, each ${B}_{{{\ell }}_{i}}(\xi )\in R$ contains a stabilizer that intersects L in at least two locations, ${{\ell }}_{i}$ and ${{\ell }}_{j}$. Thus, a single ancilla leakage could produce the error ${P}_{{{\ell }}_{i}}{P}_{{{\ell }}_{j}}$, reducing the effective code distance by one.

Now suppose we have two such stabilizer generators: ${S}_{{{\ell }}_{i}}$ corresponding to ${P}_{{{\ell }}_{i}}$, and ${S}_{{{\ell }}_{i}^{\prime} }$ corresponding to ${P}_{{{\ell }}_{i}^{\prime} }$. Then ${{\ell }}_{j}$, as an element in the support of ${S}_{{{\ell }}_{i}}$ with diameter ξ, must lie in ${B}_{{{\ell }}_{i}}(\xi )$. Similarly, ${{\ell }}_{j^{\prime} }$ must lie in ${B}_{{{\ell }}_{j}^{\prime} }$. As these balls are disjoint, ${{\ell }}_{j}\ne {{\ell }}_{j^{\prime} }$.

Thus, each ball in R corresponds to an ancilla leakage which reduces the effective distance of the code by at least one, and these reductions combine so that ${d}_{\mathrm{eff}}\leqslant d-| R| $. As $| R| \geqslant \tfrac{d}{{c}_{D}{\xi }^{D}}$, we may conclude that ${d}_{\mathrm{eff}}\leqslant \eta d$, where η is given by $1-\tfrac{1}{{c}_{D}{\xi }^{D}}$.□

Practically speaking, for most popular topological codes, the effective distance is halved. Because DP-leakage is so damaging, one might expect that any code family, when restricted to minimal overhead leakage elimination, would incur a linear distance reduction. However this is not the case: if one relaxes the practical restriction of a topological generating set, then we can manage ${d}_{\mathrm{eff}}=d-1$ in the surface code by overlapping measurement supports; see appendix D.

A more plausible solution is to use subsystem codes, in which we can measure operators that anticommute with (dressed) logical operators. Such codes are natural to consider, as their increased locality may require less relative overhead to be robust in the presence of leakage. In fact, we will see that subsystem surface codes provide a subsystem counterexample to proposition 1.

3.3. Subsystem surface codes

3.3.1. Constructions

In this section, we construct subsystem surface codes from the perspective of ensuring robustness in the presence of leakage. On a standard lattice, these codes are equivalent to those introduced in [53], and have weight-6 stabilizers that can be expressed as the product of weight-3 gauge operators in the bulk. We extend these codes to a rotated lattice, which have the same-weight bulk stabilizers and gauge operators. Opposite the subspace surface codes, the standard lattice has weight-2 boundary operators while the rotated lattice has weight-3 boundary operators.

Begin with the square lattice defining the surface code and insert a data qubit into the center of each plaquette. This triangulates the square lattice on which the surface code was initially defined, doubling the distance of the code with respect to Z-type errors. Furthermore, measuring these newly formed triangular X-type stabilizers in the presence of leakage groups the original data qubits into 'hooks', in which leakage can only recreate the hook errors defined in [41]. Unfortunately, this asymmetry between X and Z produces higher-weight, problematic hexagonal Z-type stabilizers. Measuring these larger Z-type stabilizers directly will damage the code more than measuring the stabilizers defined on the original square lattice.

The simple fix is to symmetrize the X- and Z-type operators: make both the X-plaquettes and the Z-plaquettes hexagonal. As a result, we have ${(d-1)}^{2}$ gauge degrees of freedom that we may use to measure each hexagonal stabilizer as a product of triangular gauge operators. Intuitively, as this groups the original data qubits into hooks for both X- and Z-type measurements, it should preserve the distance of the code in the presence of DP-leakage.

The price for this locality (as with many subsystem codes) is more qubits and higher-weight stabilizers, which in turn yield higher logical error rates and lower thresholds. This can be realized by relating the code capacity threshold to a phase transition of the random-bond Ising model on the honeycomb lattice along the Nishimori line [53, 74]. The resulting threshold estimate yields $p\approx 7 \% $ [75], compared to the surface code threshold estimate on a square lattice of $p\approx 11 \% $ [76].

See figure 7 for a pictoral description of these codes with boundary, which form a subsystem code family with parameters $[[3{d}^{2}-2d,1,d]]$ and ${(d-1)}^{2}$ gauge degrees of freedom. These codes inherit several nice properties from the subspace surface codes, including defect-based logical encoding, similar transversal gates, and efficient minimum-weight perfect matching decoding. Unfortunately, these codes also have significant qubit overhead per distance. For example, the smallest error-correcting code in the family forms a [[21, 1, 3]] code.

Figure 7.

Figure 7. A distance-5 subsystem surface code on a standard lattice. The X-type stabilizers are defined by the hexagonal plaquettes on the left, and similarly for Z-type stabilizers on the right. The dotted lines form the dual lattices, which represent gauge operators of opposite type. In particular, the red X-type stabilizer may be realized as the product of the two red X-type gauges, and similarly for the Z-type operators in blue. Boundaries can be assigned as weight-2 operators of the same type along opposite sides of the lattice.

Standard image High-resolution image

Fortunately, analogous to the surface code, we can rotate the lattice in order to reduce this overhead. However, unlike the rotated surface code, the boundaries are fixed by the anisotropic orientation of the stabilizers. This subsystem code family has parameters $\left[\left[\tfrac{3}{2}{d}^{2}-d+\tfrac{1}{2},1,d\right]\right]$ with $\tfrac{{(d-1)}^{2}}{2}$ gauge degrees of freedom. In particular, the smallest error-correcting code in this family forms an $[[11,1,3]]$ code with at most weight-3 check measurements, see figure 8.

Figure 8.

Figure 8. Distance 3 (top) and distance-5 (bottom) subsystem surface codes on a rotated lattice. The plaquettes on the left-side represent X-type stabilizers, while plaquettes on the right represent Z-type stabilizers. Note that XL spans the lattice from north to south, while ZL spans the lattice from east to west. Again, the dual lattices representing gauge operators of opposite type are outlined by the dotted lines.

Standard image High-resolution image

3.3.2. Ancilla leakage

Having defined the subsystem surface code families, we turn to analyzing their correlated errors in the presence of leakage. Time-correlated syndrome configurations may be treated as before, and so we may again restrict our attention to space-correlated errors. We consider ancilla leakage and data leakage separately.

Ancilla leakage is much simpler to handle in this code, as the gauge generators are all weight-3. Select any triangular X-gauge generator; by symmetry, the same analysis will apply to all other generators. Then, up to gauge transformation, any error configuration of X-type will produce an effective weight one error on the data. As data leakage in the MS-model also causes no new space-correlated error configurations, we may immediately conclude that these codes are robust to MS-leakage.

Thus, we may focus solely on DP-leakage. In particular, only Z-type error configurations occurring on X-type gauge operators (and vice versa) may produce higher-weight correlated errors. Of the three possible weight-2 configurations of Z-errors on an X-gauge operator, two are equivalent to weight one Z-errors. Thus, we need only consider two new errors in the presence of DP-leaked ancillae: the remaining weight-2 Z-error, and the weight-3 Z-error that acts on the entire triangular X-gauge operator; see figure 9.

Figure 9.

Figure 9. The two new correlated Z-errors due to ancilla DP-leakage. The dotted lines represent the gauge operators formed from the dual lattice of the Z-type stabilizers. Red dots indicate violated gauge measurements. A similar analysis applies symmetrically to every other triangular gauge operator.

Standard image High-resolution image

Note that the excitations formed from both errors in figure 9 are contained in an ${L}^{\infty }$-ball of radius one in the standard lattice, but not in the rotated lattice. Thus, neither of these errors may reduce the effective distance of the subsystem surface code in the standard lattice, while the first will halve the effective distance in the rotated lattice.

Thus, it only remains to consider data leakage. However, data leakage in the DP-model may cause error propagation between different data qubits. Furthermore, this propagation will depend on the particular gate scheduling we choose. Consequently, we relegate an accounting of data leakage and gate schedulings to appendices A and B, respectively. With proper gate timings, data leakage does not cause an effective distance reduction in these codes.

3.4. Swap-LR

One final type of leakage reduction proposed for the surface code in several works [13, 23, 31, 38, 42] is foregoing auxiliary qubits by regularly swapping the roles of data and ancilla. The idea is to continually measure and reinitialize all qubits in the lattice to ensure that leakage can persist for no more than two rounds of syndrome extraction. The mechanism for doing so is by applying SWAP gates between data and ancilla every round, see figure 10.

Figure 10.

Figure 10. The swap-LR scheme switches the role of data and ancilla, with a subspace code scheduling pictured on the left. One of the three CNOT gates forming the SWAP gate cancels with the final CNOT gate of syndrome extraction, leaving a single extra CNOT gate and no qubit overhead. On the right, the bulk of the subspace surface code can be partitioned into code qubits (shaded black or grey) and ancilla qubits (unshaded) that switch roles. Removing the shaded grey qubits yields the bulk of the subsystem surface code.

Standard image High-resolution image

This has two competing effects. On the one hand, it minimizes circuit overhead during syndrome extraction, yielding fewer potential fault locations. On the other hand, it allows leakage to persist for longer, resulting in more correlated errors. In the case of DP-leakage, the space-correlated errors it produces are also different: an ancilla leakage upon preparation depolarizes the support of its measurement, and then acts as a data leakage for one more round. These new space-correlated errors can again be handled by careful gate scheduling, at the cost of introducing additional weight-2 errors that do not damage the effective code distance; see appendices A and B.

In the case of MS-leakage, data leakage can propagate no further space-correlated errors. Thus, in the presence of MS-leakage and without using a different gate scheduling, we expect swap-LR to scale comparably to syndrome-LR. The similarities in behavior were also reported in [38].

This is precisely what we observe numerically by comparing the behaviors of DP-leakage and MS-leakage in a standard subspace surface code using swap-LR. It appears that the longer-lived leakage errors are not much more damaging, as the full-distance scaling of the code is approximately preserved in the presence of MS-leakage within the error regime we consider, see figure 11.

Figure 11.

Figure 11. A comparison of DP-leakage (top) and MS-leakage (bottom) logical error rates for the standard surface code using both syndrome-LR (dashed) and swap-LR (solid) at distances 3 (circle), 5 (cross), and 7 (triangle). As expected, we observe nearly identical scaling, and swap-LR even tends to perform better. Longer-lived leakage errors do not appear to be much more damaging in this regime, as the MS-leakage logical error rates are correctly suppressed. However, as these are subspace codes, DP-leakage reduces the effective distance.

Standard image High-resolution image

4. Leakage simulations

In the last section, we established that subsystem surface codes require less relative overhead than the subspace surface codes to realize their full effective code distance in the presence of leakage. Next, we directly compare the leakage performance of subspace and subsystem surface codes in the low-error regime. To do so, we perform Monte Carlo simulations of each code in the gate error model. Although the simplicity of syndrome-LR makes it straightforward to analyze, it does not minimize circuit-volume overhead. Consequently, we consider the least expensive leakage reduction strategy: swap-LR, which requires no qubit overhead, and importantly, may preserve the locality requirements of the qubit lattice.

We focus on error rates in the ${10}^{-5}$ to ${10}^{-3}$ range. Leakage rates spanning this range have been reported for several different qubit architectures [15, 16, 23]. However, the particular choice to center around an error-rate of ${10}^{-4}$ is motivated by spontaneous Raman scattering rates in ion traps. It has been estimated that a 200–500 $\mu {\rm{s}}$ gate experiences a spontaneous scattering event with probability ${10}^{-4}\,{\rm{t}}{\rm{o}}\,{10}^{-3}$ [31, 77], and each spontaneous scattering event can populate any state with approximately equal probability. For qubits based on clock transitions, this splits between the two computational states, as well as two leakage states formed by Zeeman splitting. We then approximate that a leakage event occurs with probability $\approx 5.0\times {10}^{-5}$ [31].

4.1. Surface codes

In the depolarizing error model, the extra overhead incurred by introducing additional gauge degrees of freedom degrades the performance of the subsystem surface code. Even well below threshold, this manifests as higher logical error rates; see figure 12.

Figure 12.

Figure 12. Sub-threshold error rates at distances 3 (circle), 5 (cross), and 7 (triangle) in the presence of depolarizing noise in both a standard (top) and rotated (bottom) lattice geometry. As expected, subspace surface codes (green, orange) consistently outperform subsystem surface codes (yellow, red), in some cases by nearly an order of magnitude within the error regime we consider.

Standard image High-resolution image

However, in the presence of leakage, the locality of the subsystem codes allow them to outperform their subspace counterparts at sufficiently low error rates. In order to probe this low-error regime, we restrict ourselves to low-distance codes; see figure 13. We observe good agreement with the expected performance of each code. Each fit takes the form ${p}_{L}\sim {p}^{{d}_{\mathrm{emp}}}$, where ${d}_{\mathrm{emp}}$ is chosen to minimize the ${\chi }^{2}$-distance. For distances 3 and 5, we observe that ${d}_{e}-0.25\lt {d}_{{\rm{emp}}}\lt {d}_{e}+0.5$, where de is the expected scaling based on syndrome-LR (see figure C1). For distance 7, the samples were drawn predominantly from error rates $\gt 3\times {10}^{-4}$. At higher rates, the distance-4 suppression of the Pauli errors are a significant factor when compared to the distance-2 suppression of leakage errors. Nonetheless, we observe the expected pronounced distance reduction, with $2\lt {d}_{\mathrm{emp}}\lt 3$ for those codes susceptible to leakage, and ${d}_{\mathrm{emp}}\gt 4$ for those robust to leakage.

Figure 13.

Figure 13. Logical error rates at distances 3 (circle), 5 (cross), and 7 (triangle) in the presence of leakage. Comparisons are between, in ascending order of qubit overhead, the rotated subspace (green), rotated subsystem (yellow), standard subspace (orange), and standard subsystem (red) surface codes using swap-LR. Each data point was recorded after at least 200 failures, with the longest simulations requiring $\approx {10}^{9}$ trials.

Standard image High-resolution image

Most importantly, for all codes that are leakage robust, we observe that ${d}_{\mathrm{emp}}\gt {d}_{e}-0.25$, giving evidence that the longer-lived leakage errors introduced during swap-LR are significantly less damaging than the space-correlated errors shared with syndrome-LR. Furthermore, we expect that error rates would improve to ${d}_{\mathrm{emp}}\approx {d}_{e}$ in this error regime given a decoder that could better handle long-range correlations, such as those based on renormalization groups [78, 79]. It is likely that these results would smooth out at higher distances, as off-by-one errors have a significant effect on the low distance scaling.

This establishes that certain subsystem codes relatively outperform subspace codes (at low distance), in the sense that same-distance, same-geometry subsystem codes yield better performance around $p\lessapprox 2.0\times {10}^{-3}$ for MS-leakage in a rotated geometry and $p\lessapprox 0.75\times {10}^{-3}$ for DP-leakage in a standard geometry. This contrasts with a local depolarizing model, in which subsystem codes do not relatively outperform surface codes in any error regime.

It is vitally important to note that this does not account for the $1.75\times $ qubit overhead required for subsystem codes of the same geometry as subspace codes. So while the arguments of section 3 demonstrate that subsystem codes offer better per-qubit distance protection in the presence of leakage, this advantage manifests at much lower error rates. Although we are unable to probe these error regimes directly at higher distances where direct per-qubit comparisons can be made, we can give coarse upper bounds using heuristic estimates based on thresholds, see appendix E. We estimate that, at the very least, p must be $\lessapprox 2.5\times {10}^{-4}$ in the presence of DP-leakage and $\lessapprox 0.32\times {10}^{-4}$ in the presence of MS-leakage to potentially see a per-qubit benefit at sufficiently high distance.

4.2. Bacon–Shor codes

We next consider Bacon–Shor codes for use in the MS-leakage model. Bacon–Shor codes are subsystem codes defined on a lattice of $d\times d$ data qubits. They have $2(d-1)$ stabilizer generators, with X- and Z-type generators indexed by $1\leqslant k\leqslant d-1$ and defined by

where ${P}_{i,j}$ represents the operator P acting on the qubit in the ith row and jth column of the lattice. Then the stabilizer group is generated by the set of all ${X}_{k},{Z}_{k}$. As subsystem codes, a generating set for the gauge operators is given by

where addition is performed modulo the lattice size d. With this orientation, XL consists of X operators spanning the north–south boundaries of the lattice, while ZL consists of Z operators spanning the east–west boundaries of the lattice. In particular, Bacon–Shor codes share the same efficient data qubit scaling as rotated surface codes, forming a family of $[[{d}^{2},1,d]]$ codes.

Although there is additional ancilla overhead, this gives us two code families to compare more closely, namely the Bacon–Shor and rotated subspace surface codes in an MS-leakage model and with bare-ancilla extraction. The former sacrifices syndrome information to perform localized checks, with weight two gauge operators ensuring that every space-correlated error due to MS-leakage has weight one. The latter sacrifices locality for additional syndrome information, increasing the number of high-weight correctable errors while introducing other damaging correlated errors in the process. The result of this tradeoff is that, at low distances, Bacon–Shor codes yield a per-qubit error-corrective advantage in the presence of MS-leakage at reasonable error rates, ranging from 10−4 to 10−3; see figure 14.

Figure 14.

Figure 14. A comparison of distance 3 (circle), 5 (cross), and 7 (triangle) rotated subspace surface codes (green) with Bacon–Shor codes (blue). Here we consider both ${p}_{{\ell }}={p}_{d}$ (top) as well as ${p}_{{\ell }}=0.1{p}_{d}$ (bottom), which may be applicable to different architectures. In both cases, Bacon–Shor codes begin to demonstrate an advantage at errors rates $\lessapprox {10}^{-3}$ at distance 3. Furthermore, this advantage persists within reasonable leakage regimes through distance 7. Each fit is chosen empirically to be of the form ${{Ap}}^{{d}_{e}}+{{Bp}}^{{d}_{e}+1}$ due to contributions from Pauli errors, where de is the minimum number of faults that may cause a logical error and $A,B$ are the fitting parameters.

Standard image High-resolution image

At low distance, this advantage manifests at error rates as high as $p\lessapprox 1.2\times {10}^{-3}$. Even when the leakage rate is an order-of-magnitude less than the depolarizing rate, the distance 3 Bacon–Shor code begins to outperform the corresponding surface code at error rates as high as $p\lessapprox 5.0\times {10}^{-4}$. Of course, these gains are more pronounced at higher leakage-to-depolarizing ratios. At lower error rates, one can observe the damage caused by leakage as the logical error suppression tapers off.

However, as the number of stabilizer checks in Bacon–Shor codes scale sublinearly in the lattice size, the family does not exhibit a threshold. Consequently, their advantage cannot persist at higher distances, although they can yield significant error-suppression which may suffice for the desired memory time [57, 58]. Coupled with simplified preparation and potential benefits against other correlated noise sources, this gives evidence that Bacon–Shor codes might prove advantageous in near-term fault-tolerance experiments or in low to intermediate distance error protection, within certain noise models [65].

4.3. Surface codes with unverified cat states

Thus far, we have focused on codes defined on square qubit lattices with bare-ancilla syndrome extraction. While this might be preferable architecturally, we may relax this requirement to consider slightly larger ancilla states. Unlike the preceding results, which can be extended straightforwardly to the presence of correlated leakage by passing to syndrome-LR, this relies unavoidably on an independent leakage model. For some platforms, such as certain quantum dot architectures, these correlated leakage events must be handled. For others, we can leverage the independence assumption to combine the locality of the Bacon–Shor code with the performance of the surface code by encoding the gauge-fix directly into the ancilla. This amounts to performing the simplest Shor-style extraction, without verification or decoding.

Typical mechanisms for fault-tolerance involve preparing and verifying entangled ancilla states [8083] or the use of flags [84, 85]. Unfortunately, as a leakage event can manifest many Pauli errors, it can often fool such verification schemes. Although this can make typical fault-tolerant circuits more difficult to engineer [11], the circuit locality of Shor-style syndrome extraction may provide a direct benefit when leakage events occur independently; see figure 15.

Figure 15.

Figure 15. Syndrome extraction on the rotated surface code with unverified two-qubit cat states. The usual circuit involves decoding the ancilla [86], with dashed gates omitted as verification can be fooled by leakage. On the right is a distance-5 code with black dots representing data qubits, red dots representing X-type ancilla qubits, and blue dots representing Z-type ancilla qubits. Each ancilla extracts the syndrome from its two nearest neighbor data qubits, with green connections representing longer-range ancilla entangling operations.

Standard image High-resolution image

Note that without entangling the ancilla, each measurement simply corresponds to measuring the gauge generators of the Bacon–Shor code. In this sense, entangling the ancilla encodes a gauge-fix of the Bacon–Shor code to the surface code [46], while robustness to MS-leakage is unaffected by this leading operation. One must at least pass to four-qubit cat states in the presence of DP-leakage, as any pair of qubits in the stabilizer supports two qubits of a minimal distance logical operator.

Compared to the previous fault-tolerance approach of [38], in which leakage reduction is performed after each gate, we observe an orders-of-magnitude improvement in performance; see figure 16. Similar to the Bacon–Shor code, using two-qubit cat state extraction outperforms bare-ancilla extraction below error rates $\approx {10}^{-3}$. However, as this family exhibits a threshold, we would expect that this advantage is better preserved at higher distances. When compared directly to Bacon–Shor codes, two-qubit cat state extraction of the surface code performs marginally worse at distance 3, and then begins to outperform.

Figure 16.

Figure 16. Error-rate comparison in the presence of MS-leakage on the rotated surface code. Green represents swap-LR, the dotted black line represents the gate-LR proposed in [38], and the dashed black line represents intermediate-LR (see appendix C). The solid black line represents swap-LR combined with unverified two-qubit cat state extraction.

Standard image High-resolution image

However, in the presence of correlated leakage events, the preparation of an entangled ancilla state may be badly corrupted. In this case, two-qubit cat states do not preserve the effective distance, highlighting that the more damaging the error model, the more valuable the subsystem locality. The resulting comparison between cat state extraction of the surface code and bare-ancilla extraction of the Bacon–Shor code in the presence correlated leakage events roughly mirrors the comparison between bare-ancilla extraction of both codes with independent leakage events; see appendix F.

5. Conclusions

In this work, we have highlighted an intrinsic error-corrective benefit of subsystem codes in the presence of leakage. This is because leakage naturally manifests uncontrolled 'hook' errors, whose damage is limited by local measurements. We have quantified several crossover points below which these correlated errors are more damaging than the entropic effects that are well-handled by the subspace surface code. These error rates range between approximately ${10}^{-5}\,{\rm{a}}{\rm{n}}{\rm{d}}\,{10}^{-3}$ at the low distances we have considered, and may be relevant to limiting error models in certain architectures, such as spontaneous scattering in ion traps.

It is important to note that some of these advantages are significant, while others are modest. For the geometrically local codes we have considered, subsystem error-correction only provides a benefit subject to certain code constraints or particularly damaging leakage models. Even here, these benefits can only be observed at low error-rates that are only just being broached by current technologies [16, 87, 88]. However, there are certain architectures that could have even more harmful leakage dynamics or even higher leakage rates [37, 89]; in such cases, we would expect the subsystem code advantage to increase.

More generally, this work illustrates a potential application for subsystem codes which do not yield an immediate error-corrective advantage in independent depolarizing models. As a generalization of subspace codes, subsystem codes offer a broad design space, with a variety of interesting constructions including Bravyi–Bacon–Shor codes [48, 90], the five-squares code [56], subsystem color codes [54], and others [9194]. Leakage is a natural error-model in which the subsystem structure of these codes offers a direct error-corrective benefit when compared with any gauge-fix, past architectural and parallelizability considerations. We would expect similar advantages in other non-Markovian error models, where the relevant baths are local and determined by the qubit connectivity [66].

For more immediate practical applications, we have given evidence that leakage can be less damaging in certain architectures. This occurs in a non-interactive model of leakage, and can be applied to Mølmer–Sørensen gates in ion-traps. In practice, one should leverage all of the specificities of their architecture to handle leakage events optimally. For systems that suffer from very damaging leakage, we have shown that Bacon–Shor or subsystem surface codes may be preferable, particularly in near-term fault-tolerance experiments at low distances. Finally, we have shown that in certain leakage models, using unverified two-qubit cat state extraction on the surface code may be preferable at error-rates $\lessapprox {10}^{-3}$ at the same-distance. This yields an order-of-magnitude improvement over existing fault-tolerant proposals for gate-by-gate leakage reduction [38].

There are several extensions to this work worth consideration. Most immediately, each of these codes and leakage reduction strategies should be compared on a qubit-to-qubit basis, given the constraints of a particular architecture and leakage model. To do this, we would have to probe the low-error regime of higher-distance codes. Generalized path-counting [71, 95] or splitting methods [96] may be better suited to the task than Monte Carlo simulations. Ultimately, we are interested in performing very long computations, and so understanding code behavior at high distances and low error-rates is key.

Additionally, one could expand the search for leakage resilient codes past geometrically local codes, which are bounded asymptotically by $d\,=\,O(\sqrt{n})$, and consider more general low-density parity check codes [90]. Even among geometrically local codes, one could consider more qubit-efficient encodings, such as the $\mathrm{4.8.8}$ color codes [97] or triangle codes [86]. One could also try to generalize fault-tolerant syndrome extraction to account for leakage events [11]. However, this will also introduce additional overhead, while preparation and verification of the required resource states will prove more difficult depending on the leakage model. Along the same line, one could investigate concatenated codes for non-local architectures. However, leakage will further exacerbate the low thresholds inherent to concatenated codes.

One could also look for improvements using more advanced decoders or leakage detection [38]. We have restricted our attention to minimum-weight perfect matching, which only captures edge-like correlated errors in the decoder graph. Renormalization group [78, 79] or machine-learning methods [98, 99] could better handle higher-weight correlated error patterns. These decoders might also utilize the extra information that is gained from auxiliary qubit measurement, which is wasted in a simple minimum-weight perfect matching scheme, and could lead to improved thresholds.

Finally, one might consider subsystem codes in the presence of other correlated noise models [66, 100], particularly those that are determined by the lattice geometry or gate connectivity, such as crosstalk. In these cases, local subsystem error-correction may provide benefits for handling correlated errors that lie outside the scope of independent depolarizing noise.

Acknowledgments

The authors thank Andrew Cross and Martin Suchara for providing the toric code simulator from which the other simulators were built, with permission from IBM. They additionally thank Pavithran Iyer, Aleksander Kubica, Yukai Wu, and Ted Yoder for helpful comments and discussions. Finally, they acknowledge the Duke Computing Cluster for providing the computational resources for simulations. This research was supported in part by NSF (1717523), ODNI/IARPA LogiQ program (W911NF-10-1-0231), ARO MURI (W911NF-16-1-0349), and EPiQC—an NSF Expedition in Computing (1730104).

Appendix A.: Data leakage

In this section, we detail the requirements for a gate scheduling to ensure that data leakage in the DP-model does not damage the effective distance of subsystem surface codes. Unfortunately, gate schedulings for subsystem codes are more complicated than for subspace codes. This is because measuring anticommuting gauge operators may randomize their syndrome outcomes. Consequently, interleaving and parallelizing syndrome extraction must be performed carefully. For some architectures like ion-traps, errors are dominated by gate application, with memory errors occurring orders-of-magnitude less frequently [101]. In these cases, parallelization becomes relatively less important. However, for other architectures like superconductors with gates limited by memory error, this parallelization is vital [43].

A.1. Syndrome-LR

There are two properties that we want our syndrome extraction to satisfy. The first is that it is correct: its outcomes correctly reflect Pauli errors on the data in the ideal case (i.e. there is no randomization due to measuring anticommuting operators consecutively). The second property we require is that two triangular gauge operators of the same type overlapping on any qubit cannot both interact with that qubit in the second time-step after their initialization. This requirement removes bad correlated errors due to data leakage. We call a syndrome extraction scheduling satisfying this property good.

In order to avoid the pitfall of measuring anticommuting operators, the simplest syndrome extraction is to first measure all X-type gauges in parallel, then measure all Z-type gauges in parallel. In [53], a rolling syndrome extraction was defined that ensured that no qubit was idle at any time-step during the syndrome extraction of the subsystem surface code. To simplify the discussion, we perform the simplest syndrome extraction in our gate error model without idling errors. We further motivate this simplification by detailing a fully-parallelized (on data qubits) good and correct syndrome extraction scheme for subsystem surface codes using syndrome-LR in appendix B.

It remains to argue that data leakage may not cause damaging correlated errors in a good and correct syndrome extraction scheme. Note that if an ancilla interacts with a data leakage in the third time-step after initialization, then it interacts with no other data qubits after that point and cannot spread a correlated error. If an ancilla interacts with a data leakage in the first time-step after initialization, then the resulting correlated error is equivalent to an error on the leaked qubit, up to gauge transformation. Thus, two-qubit errors due to data leakage may only occur in triangular gauge operators which both interact with the leakage in the second time-step after initialization.

Furthermore, each gauge operator may only propagate Pauli errors of the same type. As error-correction is performed independently, it suffices for leakage robustness that no two gauge operators of the same type both introduce correlated errors due to a single leakage. Thus, goodness of syndrome extraction ensures that any error of either X- or Z-type caused by a single data leakage is contained in one triangular gauge operator of that type. As this falls into the ancilla leakage model, we may conclude that the effective distance of the code is preserved. See figure A1 for the damaging data leakage that might occur in a bad syndrome extraction scheduling. This completes the analysis for subsystem surface codes using syndrome-LR.

Figure A1.

Figure A1. Potential data leakage errors. The left-hand error is equivalent to an ancilla leakage, producing the red excitations if it depolarizes to I and the red plus yellow excitations if it depolarizes to X, and so it cannot reduce the effective distance. In contrast, if the right-hand leakage error depolarizes to the identity, it causes a separated pair of excitations, reducing the distance. With good syndrome extraction, only one of the X-errors on the right-hand side (caused by propagation of an X-error to an ancilla) may occur.

Standard image High-resolution image

A.2. Swap-LR

Similar to data leakage using syndrome-LR, we must ensure that errors of either X- or Z-type produced by a single data leakage event are contained in the support of a single gauge operator. In the case of syndrome-LR, long-range correlated errors could occur when two triangular gauge operators of the same type both interact with a data leakage in time-step two. In the case of swap-LR, because an ancilla leakage swaps places with a data qubit in time-step three, long range correlated errors could occur when two triangular gauge operators interact with a data qubit in any combination of time-steps two and three.

Fortunately, figure B5 demonstrates a gate scheduling that avoids these distance-reducing correlated errors. It includes a single long-range correlated error that is not too damaging, occurring when a data qubit occupying the short-side of a hexagonal stabilizer leaks; see the left-hand side of figure B1.

Appendix B.: Gate schedulings

B.1. Simulation details

We follow the simulation technique in [38]: for a code of distance-d, we simulate d time-steps of faulty syndrome extraction. Then, we run an additional round of perfect syndrome extraction, converting any remaining leakage errors to depolarizing errors in the process. As we consider the simulated round of error-correction as occurring during a longer computation, we associate a leading probability of leakage according to the lifetime of each data qubit at its initialization. Each simulation ran until at least 200 failures were recorded, and for those simulations that required less than $5\times {10}^{8}$ trials, until at least 1000 failures were recorded.

For all simulations, we use the standard minimum-weight perfect matching decoder. The edge weights are generated by iterating over all single Pauli faults, and associating a relative probability p to each edge of the decoder graph. The weight of the edge is then defined to be $-\mathrm{ln}(p)$. Boundary conditions are handled according to the techniques of [102].

Note that we do not iterate over leakage events as they may cause fault configurations that result in hyperedges. Furthermore, for codes that experience a distance reduction in the presence of leakage, we would have to insert edges lying outside a fundamental cell. Thus, these edge weights are suboptimal. However, to counteract this, we add edges corresponding to known single-time-step leakage fault configurations occurring with probabilities linear in p. We heuristically weight these as four times the maximum edge probability generated by Pauli configurations.

B.2. Timings

For the subspace surface code, we use different but standard gate schedulings depending on the lattice geometry, see figure B1. Note that the decoder graph is always populated by more edges than in the independent depolarizing model, corresponding to the additional correlated errors that may be caused by leakage.

Figure B1.

Figure B1. Scheduling for the standard (left) and rotated (right) subspace surface codes. On the rotated lattice, the left plaquette represents an X-type stabilizer, while the right represents a Z-type stabilizer. These are chosen so that hook errors occur parallel to the boundary of the same type: in this case, an X-type north–south boundary and a Z-type east–west boundary.

Standard image High-resolution image

For the subsystem surface code, we use the trivial scheduling; see figure B2. It is straightforward to check that the syndrome extraction satisfies the goodness property. For correctness, as the X- and Z-type syndromes are extracted serially, there can be no randomization due to anticommuting operators.

Figure B2.

Figure B2. Trivial scheduling for the subsystem code. Dark circles are code qubits, red silhouetted circles are X-ancilla, and blue silhouetted circles are Z-ancilla. The numbers refer to CNOT locations in a particular time-step, except for those in the center that denote measurement. Full syndrome extraction proceeds in seven time-steps (not including ancilla preparation).

Standard image High-resolution image

However, we demonstrate that there also exists a good and correct syndrome extraction scheme that is fully-parallelized, i.e. there are no idling data qubits in any time-step. For this, we will need a lemma from [53] giving criteria on when a syndrome extraction scheduling is correct.

(Lemma 1, [53]).

Lemma B.1 A fully-parallelized schedule of CNOTs is correct if for any pair of X-type and Z-type triangular gauge operators, at least one of the following is true:

  • (i)  
    the two operators are disjoint;
  • (ii)  
    the two operators are measured two rounds apart;
  • (iii)  
    the last gate of the operator measured first in the pair commutes with the first gate of the operator measured last in the pair.

For proof, see [53]. One can show that any scheduling which is translation invariant with respect to a single pair of overlapping hexagonal stabilizers, as shown in figure B3, cannot simultaneously satisfy lemma B.1 and satisfy goodness.

However, there exist such schedulings that are translation invariant on adjacent pairs of overlapping hexagonal stabilizers, and such a scheduling suffices for odd distance codes with boundary. See figure B3 for an example, whose goodness can be checked and which satisfies the hypotheses of lemma B.1. This example demonstrates that there exist fully-parallelized syndrome extractions that are both good and correct. Note that, although there are no idling data qubits, there will be (few) idling time-steps for the ancilla qubits as data qubits are swapped with auxiliary qubits.

Figure B3.

Figure B3. Fully-parallelized good and correct scheduling. Rolling syndrome extraction is performed over four different time-steps, with the middle numbers indicating when measurements are occurring. A full scheduling can be realized as translations of this fundamental cell. In the presence of leakage, one must insert swapping of the data with auxiliary qubits into the scheduling.

Standard image High-resolution image

Finally, for a swap-LR scheme defined on subsystem surface codes, long-range correlated errors may occur when triangular gauge operators of the same type interact with a data qubit in any combination of time-steps two and three. This can be avoided for all but one data qubit (up to translation-symmetry), with every data qubit swapping with some ancilla in time-step three. If that one data qubit is chosen properly, these long-range correlated errors do not damage the effective distance of the code. See figure B4 for the scheduling and the left-hand side of figure B1 for the benign correlated error.

Figure B4.

Figure B4. Subsystem code scheduling for swap-LR. The last gate in any triangular gauge operator swaps data and ancilla, except for the northeast ancilla qubit. The northwest/southeast data qubit may cause a long-range X-type correlated error as it interacts with two ancilla qubits in time-step two. However, unlike the damaging data leakage on the right-hand side of figure B1, these correlated errors may also be formed by a single Z-type ancilla leakage, up to gauge transformation. By our classification of ancilla leakage errors, they cannot reduce the effective distance of the code.

Standard image High-resolution image

Appendix C.: Auxiliary qubit leakage reduction

There are many advantages to using minimal leakage reduction; for example, preserving the connectivity of the qubit lattice. However, one could also consider strategies which involve many auxiliary qubits and LRUs. We detail two such additional leakage elimination methods here.

  • (i)  
    Gate leakage reduction (gate-LR) swaps the qubits involved in each gate with auxiliary qubits immediately after the gate is applied. A single leakage may only persist for a single time-step, effectively converting the leakage model to a depolarizing model at the cost of significant overhead.
  • (ii)  
    Intermediate leakage reduction (int-LR) applies syndrome-LR, as well as swapping with an auxiliary qubit after the second CNOT in each syndrome extraction. For subspace surface codes, it swaps both the data and the ancilla with auxiliary qubits, while for subsystem surface codes, it swaps only the ancilla.

Gate-LR was studied in [38] as a means of retaining the effective code distance. However, its tremendous overhead caused much lower thresholds, around ${p}_{\mathrm{thr}}=1.6\times {10}^{-3}$ when translated to our noise model. The higher thresholds of low-overhead subsystem code leakage reduction indicate better performance given that both preserve the effective distance of the code, but also depend on the assigned cost of the required auxiliary qubits. By direct comparison, Bacon–Shor codes also perform better at low distances.

Intermediate leakage reduction is a more intermittent leakage elimination strategy, which can preserve the effective code distance with lower overhead but allows for more correlated errors compared to gate-LR. We summarize these relationships in figure C1, which measures the circuit-volume overhead of each strategy. Note that we do not include the extra $\theta (d)$ qubits required on the boundary of the qubit lattice in order to preserve the connectivity using swap-LR.

Figure C1.

Figure C1. A summary of the overall qubit requirements for different leakage reduction strategies, where d is the distance of the code. Red represents an effective distance reduction in both MS- and DP-leakage, yellow in only DP-leakage, green in neither, while grey indicates no threshold. Motivated by qubit timings in ion-traps [101], each cell indicates the total qubit count (including data and ancilla) required in a single round of syndrome extraction. The effective distance for swap-LR is estimated numerically rather than analytically, with auxiliary qubits included for syndrome-LR.

Standard image High-resolution image

Appendix D.: Non-topological counterexample

Here we illustrate a non-topological counterexample to proposition 1. It reflects a counter-intuitive point in the proof: the upper bound on ${d}_{\mathrm{eff}}$ grows tighter as the diameter of the stabilizer decreases. One would expect that larger stabilizers would give rise to larger correlated errors, and thus would give tighter upper bounds on ${d}_{\mathrm{eff}}$.

However, large overlapping stabilizer generators may prevent bad correlated errors from combining. In fact, by relaxing the restriction of a topological generating set, we can manage ${d}_{\mathrm{eff}}=d-1$ in the surface code when decoding is performed carefully [46]. If we label each plaquette i rows from the top and j columns from the left as Pij, and similarly for vertices Vij, then this generating set may be realized as

Still, this is not a practical strategy as it requires measuring long-range stabilizers that scale with the lattice size, see figure D1. While this will unreasonably degrade code performance, it demonstrates that the direction of the upper bound and the hypotheses of proposition 1 are necessary.

Figure D1.

Figure D1. A non-topological generating set that only reduces the effective code distance by one. Each X-stabilizer is measured as an increasing product of plaquettes along each row, and each Z-stabilizer is measured as an increasing product of dual plaquettes along each column. Note, for example, that a single leakage may produce an X-type error that spans the dual lattice from north to south, but this cannot produce a logical error.

Standard image High-resolution image

Appendix E.: Threshold estimates

In this section, we directly compare subsystem surface codes to subspace surface codes, while accounting for the $1.75\times $ qubit overhead required for subsystem codes of the same distance and lattice geometry. Although we have focused on the low p limit at low distances to remain simulable, we are more interested in the error-suppression at fixed p for increasing d. Up to a polynomial pre-factor, we can express ${p}_{L}(p,d)\sim \exp (-\alpha (p)d)$. Determining the p for which ${\alpha }_{\mathrm{subsystem}}(p)\gt {\alpha }_{\mathrm{subspace}}(p)$, optimized with respect to both the qubit overhead and geometry, will determine the error regime in which subsystem surface codes absolutely outperform subspace surface codes.

Unfortunately, in the low-error regime we consider, it is difficult to obtain accurate error estimates at even intermediate distances, and low distance estimates suffer from finite-size effects. However, we can make heuristic estimates via the high-distance fit [43]

Equation (E1)

where ${p}_{\mathrm{thr}}$ is the threshold and de is the estimated scaling, based on the near-fitting of figure 13 to the minimum number of faults required to produce a logical error using syndrome-LR. We can estimate the thresholds accurately as they occupy a higher-error regime, see figure E1.

Figure E1.

Figure E1. Thresholds simulated from distance $9,13,$ and 17 lattices with periodic boundaries to minimize finite-size effects. The left column and right column represent DP- and MS-leakage, respectively, while the top row and bottom row represent subsystem and subspace codes, respectively. Each data point was obtained from $5\times {10}^{4}\,{\rm{t}}{\rm{o}}\,5\times {10}^{5}$ trials, with distance-9 lattices still exhibiting some finite-size effects.

Standard image High-resolution image

Given thresholds ${p}_{g},{p}_{s}$ for a subsystem and subspace code, respectively, and normalizing qubit-overhead factors ${\gamma }_{g},{\gamma }_{s}$ taken from table 1, we can upper-bound the error rate ${p}_{* }$ below which subsystem surface codes outperform subspace surface codes at sufficiently high distance from equation (E1),

This estimate yields ${p}_{* }^{{DP}}\approx 2.5\times {10}^{-4}$ in the presence of DP-leakage and ${p}_{* }^{{MS}}\approx 0.32\times {10}^{-4}$ in the presence of MS-leakage at sufficiently high distance. Given table 1, these ${p}_{* }$ are obtained by comparing the optimal subsystem and subspace code choices for each leakage type. These are again the standard subsystem codes and rotated subspace codes in the presence of DP-leakage, and the rotated subsystem codes and standard subspace codes in the presence of MS-leakage.

Generally, fitting formulas like equation (E1) tend to underestimate the logical error rate at finite p [96]. This, coupled with the observation that Pauli errors contribute more heavily at near-threshold errors rates, makes it likely that these fits already favor the subsystem surface codes, and hence only provide a heuristic upper-bound.

It is worth noting that our simulations actually demonstrate better low-rate scaling with rotated subsystem codes than with standard subspace codes in the presence of MS-leakage, somewhat counteracting the overestimation of ${p}_{* }^{{MS}}$. Further numerics are required to generate a more accurate estimate based on empirical scalings at higher distances.

Appendix F.: Correlated leakage

For certain architectures like quantum dots, correlated leakage events may occur simultaneously on the full support of a two-qubit gate. In this case, leakage robustness follows directly from considering correlated error patterns using swap-LR, where instead of leakage occurring on ancilla and data qubits in consecutive rounds, they occur simultaneously in the same round. Then the same set of space-correlated errors will occur as long as one instead uses syndrome-LR. Note that if one uses swap-LR directly, then in the case of MS-leakage, we may depolarize two qubits in the support of the measurement: the data qubit that is leaked and the data qubit that is swapped with the leaked ancilla.

We compare Bacon–Shor codes and surface codes with two-qubit cat state extraction in the presence of correlated leakage. Here, preparation of the cat state itself may cause a damaging error. The two codes display similar relative behavior when using bare-ancilla extraction with independent leakage, see figure F1. By the preceding argument, subsystem surface codes should also achieve the correct logical error suppression.

Figure F1.

Figure F1. Distance 3 (circle), 5 (cross), and 7 (triangle) rotated subspace surface codes (green) compared with Bacon–Shor codes (blue) in the presence of correlated MS-leakage. Bacon–Shor codes begin to demonstrate an advantage at errors rates $\lessapprox {10}^{-3}$ at distance 3. This advantage drops off quickly at higher distances.

Standard image High-resolution image

Footnotes

  • The twirling approximation allows us to more efficiently simulate leakage in a Pauli model, but can also change the dynamics. For example, if two leaked data qubits are measured by a stabilizer in the Clifford model, the outcome is anti-correlated with the parity of anticommuting errors on the sealed data qubits, rather than random. This is equivalent to a measurement error at a much lower error rate than native measurement errors. The divergence between these two models is small, and in most cases, Clifford errors do not interact so that the Pauli-twirl gives similar results.

  • For standard geometries, we approximate the code with boundary by a code with periodic boundaries for threshold computations, but only preserve a single logical qubit.

Please wait… references are loading.