Top arXiv papers

sign in to customize
  • PDF
    We refine ideas from Knill 1996, Jones 2016, Chamberland 2020, Gidney 2023+2024, Bombin 2024, and Hirano 2024 to efficiently prepare good $|T\rangle$ states. We call our construction "magic state cultivation" because it gradually grows the size and reliability of one state. Cultivation fits inside a surface code patch and uses roughly the same number of physical gates as a lattice surgery CNOT gate of equivalent reliability. We estimate the infidelity of cultivation (from injection to idling at distance 15) using a mix of state vector simulation, stabilizer simulation, error enumeration, and Monte Carlo sampling. Compared to prior work, cultivation uses an order of magnitude fewer qubit-rounds to reach logical error rates as low as $2 \cdot 10^{-9}$ when subjected to $10^{-3}$ uniform depolarizing circuit noise. Halving the circuit noise to $5 \cdot 10^{-4}$ improves the achievable logical error rate to $4 \cdot 10^{-11}$. Cultivation's efficiency and strong response to improvements in physical noise suggest that further magic state distillation may never be needed in practice.
  • PDF
    Free-fermionic states, also known as fermionic Gaussian states, represent an important class of quantum states ubiquitous in physics. They are uniquely and efficiently described by their correlation matrix. However, in practical experiments, the correlation matrix can only be estimated with finite accuracy. This raises the question: how does the error in estimating the correlation matrix affect the trace-distance error of the state? We show that if the correlation matrix is known with an error $\varepsilon$, the trace-distance error also scales as $\varepsilon$ (and vice versa). Specifically, we provide distance bounds between (both pure and mixed) free-fermionic states in relation to their correlation matrix distance. Our analysis also extends to cases where one state may not be free-fermionic. Importantly, we leverage our preceding results to derive significant advancements in property testing and tomography of free-fermionic states. Property testing involves determining whether an unknown state is close to or far from being a free-fermionic state. We first demonstrate that any algorithm capable of testing arbitrary (possibly mixed) free-fermionic states would inevitably be inefficient. Then, we present an efficient algorithm for testing low-rank free-fermionic states. For free-fermionic state tomography, we provide improved bounds on sample complexity in the pure-state scenario, substantially improving over previous literature, and we generalize the efficient algorithm to mixed states, discussing its noise-robustness.
  • PDF
    In this work we investigate how quantum expanders (i.e. quantum channels with few Kraus operators but a large spectral gap) can be constructed from unitary designs. Concretely, we prove that a random quantum channel whose Kraus operators are independent unitaries sampled from a $2$-design measure is with high probability an optimal expander (in the sense that its spectral gap is as large as possible). More generally, we show that, if these Kraus operators are independent unitaries of the form $U^{\otimes k}$, with $U$ sampled from a $2k$-design measure, then the corresponding random quantum channel is typically an optimal $k$-copy tensor product expander, a concept introduced by Harrow and Hastings (Quant. Inf. Comput. 2009).
  • PDF
    We develop the theory of local operations and classical communication (LOCC) for bipartite quantum systems represented by commuting von Neumann algebras. Our central result is the extension of Nielsen's Theorem to arbitrary factors. As in the matrix algebra case, the LOCC ordering of bipartite pure states is connected to the majorization of their restrictions. Our theorem implies that, in a bipartite system modeled by commuting factors in Haag duality, a) all states have infinite one-shot entanglement if and only if the local factors are not of type I, b) type III factors are characterized by LOCC transitions of arbitrary precision between any two pure states, and c) the latter holds even without classical communication for type III$_{1}$ factors. In the case of semifinite factors, the usual construction of entanglement monotones carries over using majorization theory. In the appendix, we provide a self-contained treatment of majorization on semifinite von Neumann algebras and $\sigma$-finite measure spaces.
  • PDF
    We consider an open quantum system undergoing Markovian dynamics, the latter being modelled by a discrete-time quantum Markov semigroup $\{\Phi^n\}_{n \in {\mathbb{N}}}$, resulting from the action of sequential uses of a quantum channel $\Phi$, with $n \in {\mathbb{N}}$ being the discrete time parameter. We find upper and lower bounds on the one-shot $\epsilon$-error information transmission capacities of $\Phi^n$ for a finite time $n\in \mathbb{N}$ and $\epsilon \in [0,1)$ in terms of the structure of the peripheral space of the channel $\Phi$. We consider transmission of $(i)$ classical information (both in the unassisted and entanglement-assisted settings); $(ii)$ quantum information and $(iii)$ private classical information.
  • PDF
    Exactly solvable models in quantum many body dynamics provide valuable insights into many interesting physical phenomena, and serve as platforms to rigorously investigate fundamental theoretical questions. Nevertheless, they are extremely rare and existing solvable models and solution techniques have serious limitations. In this paper we introduce a new family of exactly solvable unitary circuits which model quantum many body dynamics in discrete space and time. Unlike many previous solvable models, one can exactly compute the full quantum dynamics initialized from any matrix product state in this new family of models. The time evolution of local observables and correlations, the linear growth of Renyi entanglement entropy, spatiotemporal correlations, and outof-time-order correlations are all exactly computable. A key property of these models enabling the exact solution is that any time evolved local operator is an exact matrix product operator with finite bond dimension, even at arbitrarily long time, which we prove using the underlying (weak) Hopf algebra structure along with tensor network techniques. We lay down the general framework for the construction and solution of this family of models, and give several explicit examples. In particular, we study in detail a model constructed out of a weak Hopf algebra that is very close to a floquet version of the PXP model, and the exact results we obtain may shed light on the phenomenon of quantum many body scars, and more generally, floquet quantum dynamics in constrained systems.
  • PDF
    Cat qubits, a type of bosonic qubit encoded in a harmonic oscillator, can exhibit an exponential noise bias against bit-flip errors with increasing mean photon number. Here, we focus on cat qubits stabilized by two-photon dissipation, where pairs of photons are added and removed from a harmonic oscillator by an auxiliary, lossy buffer mode. This process requires a large loss rate and strong nonlinearities of the buffer mode that must not degrade the coherence and linearity of the oscillator. In this work, we show how to overcome this challenge by coloring the loss environment of the buffer mode with a multi-pole filter and optimizing the circuit to take into account additional inductances in the buffer mode. Using these techniques, we achieve near-ideal enhancement of cat-qubit bit-flip times with increasing photon number, reaching over $0.1$ seconds with a mean photon number of only $4$. Concurrently, our cat qubit remains highly phase coherent, with phase-flip times corresponding to an effective lifetime of $T_{1,\text{eff}} \simeq 70$ $\mu$s, comparable with the bare oscillator lifetime. We achieve this performance even in the presence of an ancilla transmon, used for reading out the cat qubit states, by engineering a tunable oscillator-ancilla dispersive coupling. Furthermore, the low nonlinearity of the harmonic oscillator mode allows us to perform pulsed cat-qubit stabilization, an important control primitive, where the stabilization can remain off for a significant fraction (e.g., two thirds) of a $3~\mathrm{\mu s}$ cycle without degrading bit-flip times. These advances are important for the realization of scalable error-correction with cat qubits, where large noise bias and low phase-flip error rate enable the use of hardware-efficient outer error-correcting codes.
  • PDF
    Discrete models of holographic dualities, typically modeled by tensor networks on hyperbolic tilings, produce quantum states with a characteristic quasiperiodic disorder not present in continuum holography. In this work, we study the behavior of XXZ spin chains with such symmetries, showing that lessons learned from previous non-interacting (matchgate) tensor networks generalize to more generic Hamiltonians under holographic disorder: While the disorder breaks translation invariance, site-averaged correlations and entanglement of the disorder-free critical phase are preserved at a plateau of nonzero disorder even at large system sizes. In particular, we show numerically that the entanglement entropy curves in this disordered phase follow the expected scaling of a conformal field theory (CFT) in the continuum limit. This property is shown to be non-generic for other types of quasiperiodic disorder, only appearing when our boundary disorder ansatz is described by a "dual" bulk hyperbolic tiling. Our results therefore suggest the existence of a whole class of critical phases whose symmetries are derived from models of discrete holography.
  • PDF
    Topological orders in 2+1d are spontaneous symmetry-breaking (SSB) phases of 1-form symmetries in pure states. The notion of symmetry is further enriched in the context of mixed states, where a symmetry can be either ``strong" or ``weak". In this work, we apply a Rényi-2 version of the proposed equivalence relation in [Sang, Lessa, Mong, Grover, Wang, & Hsieh, to appear] on density matrices that is slightly finer than two-way channel connectivity. This equivalence relation distinguishes general 1-form strong-to-weak SSB (SW-SSB) states from phases containing pure states, and therefore labels SW-SSB states as ``intrinsically mixed". According to our equivalence relation, two states are equivalent if and only if they are connected to each other by finite Lindbladian evolution that maintains continuously varying, finite Rényi-2 Markov length. We then examine a natural setting for finding such density matrices: disordered ensembles. Specifically, we study the toric code with various types of disorders and show that in each case, the ensemble of ground states corresponding to different disorder realizations form a density matrix with different strong and weak SSB patterns of 1-form symmetries, including SW-SSB. Furthermore we show by perturbative calculations that these disordered ensembles form stable ``phases" in the sense that they exist over a finite parameter range, according to our equivalence relation.
  • PDF
    We consider the dynamics of generic chaotic quantum many-body systems with no conservation laws, subject to weak bulk dissipation. It was recently observed [T. Mori, arXiv:2311.10304] that the generator of these dissipative dynamics, a quantum channel $\mathcal{E}$, retains a nonzero gap as the dissipation strength $\gamma \to 0$ if the thermodynamic limit is taken first. We use a hydrodynamic description of operator spreading in the presence of dissipation to estimate the gap of $\mathcal{E}$ as $\gamma \to 0$; to calculate the operator-size distribution of the low-lying eigenmodes of $\mathcal{E}$; and to relate the gap to the long-time decay rates of autocorrelation functions under unitary dynamics. We provide a microscopic derivation of this hydrodynamic perspective for random unitary circuits. We argue that the gap in the $\gamma \to 0$ limit can change nonanalytically as one tunes the parameters of the unitary dynamics.
  • PDF
    Codesign, an integral part of computer architecture referring to the information interaction in hardware-software stack, is able to boost the algorithm mapping and execution in the computer hardware. This well applies to the noisy intermediate-scale quantum era, where quantum algorithms and quantum processors both need to be shaped to allow for advantages in experimental implementations. The state-of-the-art quantum adiabatic optimization algorithm faces challenges for scaling up, where the deteriorating optimization performance is not necessarily alleviated by increasing the circuit depth given the noise in the hardware. The counterdiabatic term can be introduced to accelerate the convergence, but decomposing the unitary operator corresponding to the counterdiabatic terms into one and two-qubit gates may add additional burden to the digital circuit depth. In this work, we focus on the counterdiabatic protocol with a codesigned approach to implement this algorithm on a photonic quantum processor. The tunable Mach-Zehnder interferometer mesh provides rich programmable parameters for local and global manipulation, making it able to perform arbitrary unitary evolutions. Accordingly, we directly implement the unitary operation associated to the counterdiabatic quantum optimization on our processor without prior digitization. Furthermore, we develop and implement an optimized counterdiabatic method by tackling the higher-order many-body interaction terms. Moreover, we benchmark the performance in the case of factorization, by comparing the final success probability and the convergence speed. In conclusion, we experimentally demonstrate the advantages of a codesigned mapping of counterdiabatic quantum dynamics for quantum computing on photonic platforms.
  • PDF
    While commonly used entanglement criteria for continuous variable systems are based on quadrature measurements, here we study entanglement detection from measurements of the Wigner function. These are routinely performed in platforms such as trapped ions and circuit QED, where homodyne measurements are difficult to be implemented. We provide complementary criteria which we show to be tight for a variety of experimentally relevant Gaussian and non-Gaussian states. Our results show novel approaches to detect entanglement in continuous variable systems and shed light on interesting connections between known criteria and the Wigner function.
  • PDF
    Quantum computing with qudits, an extension of qubits to multiple levels, is a research field less mature than qubit-based quantum computing. However, qudits can offer some advantages over qubits, by representing information with fewer separated components. In this article, we present QuForge, a Python-based library designed to simulate quantum circuits with qudits. This library provides the necessary quantum gates for implementing quantum algorithms, tailored to any chosen qudit dimension. Built on top of differentiable frameworks, QuForge supports execution on accelerating devices such as GPUs and TPUs, significantly speeding up simulations. It also supports sparse operations, leading to a reduction in memory consumption compared to other libraries. Additionally, by constructing quantum circuits as differentiable graphs, QuForge facilitates the implementation of quantum machine learning algorithms, enhancing the capabilities and flexibility of quantum computing research.
  • PDF
    We study no-signalling correlations, defined over a quadruple of second countable compact Hausdorff spaces. Using operator-valued information channels over abstract alphabets, we define the subclasses of local, quantum spatial and quantum commuting measurable no-signalling correlations. En route, we establish measurable versions of the Stinespring's Dilation Theorem. We define values of measurable non-local games of local, quantum spatial and quantum commuting type, as well as inner versions thereof, and show how the asymptotic values of a finite non-local game can be viewed as special cases of the corresponding inner values of a measurable game, canonically associated with the given finite game.
  • PDF
    For any $\varepsilon > 0$, we prove that $k$-Dimensional Matching is hard to approximate within a factor of $k/(12 + \varepsilon)$ for large $k$ unless $\textsf{NP} \subseteq \textsf{BPP}$. Listed in Karp's 21 $\textsf{NP}$-complete problems, $k$-Dimensional Matching is a benchmark computational complexity problem which we find as a special case of many constrained optimization problems over independence systems including: $k$-Set Packing, $k$-Matroid Intersection, and Matroid $k$-Parity. For all the aforementioned problems, the best known lower bound was a $\Omega(k /\log(k))$-hardness by Hazan, Safra, and Schwartz. In contrast, state-of-the-art algorithms achieved an approximation of $O(k)$. Our result narrows down this gap to a constant and thus provides a rationale for the observed algorithmic difficulties. The crux of our result hinges on a novel approximation preserving gadget from $R$-degree bounded $k$-CSPs over alphabet size $R$ to $kR$-Dimensional Matching. Along the way, we prove that $R$-degree bounded $k$-CSPs over alphabet size $R$ are hard to approximate within a factor $\Omega_k(R)$ using known randomised sparsification methods for CSPs.
  • PDF
    Multi-distribution or collaborative learning involves learning a single predictor that works well across multiple data distributions, using samples from each during training. Recent research on multi-distribution learning, focusing on binary loss and finite VC dimension classes, has shown near-optimal sample complexity that is achieved with oracle efficient algorithms. That is, these algorithms are computationally efficient given an efficient ERM for the class. Unlike in classical PAC learning, where the optimal sample complexity is achieved with deterministic predictors, current multi-distribution learning algorithms output randomized predictors. This raises the question: can these algorithms be derandomized to produce a deterministic predictor for multiple distributions? Through a reduction to discrepancy minimization, we show that derandomizing multi-distribution learning is computationally hard, even when ERM is computationally efficient. On the positive side, we identify a structural condition enabling an efficient black-box reduction, converting existing randomized multi-distribution predictors into deterministic ones.
  • PDF
    Transmitting an entangled state over an extended distance is crucial for the development of quantum networks. Previous demonstrations of transmitting entangled photons over long distance using satellites or fibers have use entangled photon pairs generated from bulk crystal arrangements. An alternative approach would be to generate photon pairs using silicon-on-insulator (SOI) chips. Despite numerous proof-of-concept studies, no long range distribution has been achieved using this platform because of the challenge of getting sufficient off-chip brightness. We report a SOI platform that provides an off-chip entangled photon pair brightness of between 8,000 to 460,000 pairs per second. This exceeds previous reports by three orders of magnitude in brightness. The entanglement fidelity is 99.85(6)% and 97.90(3)% respectively. Measuring one photon locally, and transmitting the other over 93 km of deployed fiber (link loss of 40 dB), achieves a count rate of 132 pairs per second with an entanglement fidelity of 93.3(3)%, after solving the additional challenges of chromatic dispersion. The source can be pumped harder to enable transmission of entangled photons over 155 km of deployed fiber (link loss of 66 dB) at a rate of 0.7 pairs per second, with an entanglement fidelity of 87.6(5)%. These results demonstrate that SOI nanophotonic chips can perform competitively with bulk crystal sources and represent an important step toward building quantum networks using integrated nanophotonic platforms.
  • PDF
    The 2023 Big ANN Challenge, held at NeurIPS 2023, focused on advancing the state-of-the-art in indexing data structures and search algorithms for practical variants of Approximate Nearest Neighbor (ANN) search that reflect the growing complexity and diversity of workloads. Unlike prior challenges that emphasized scaling up classical ANN search ~\citeDBLP:conf/nips/SimhadriWADBBCH21, this competition addressed filtered search, out-of-distribution data, sparse and streaming variants of ANNS. Participants developed and submitted innovative solutions that were evaluated on new standard datasets with constrained computational resources. The results showcased significant improvements in search accuracy and efficiency over industry-standard baselines, with notable contributions from both academic and industrial teams. This paper summarizes the competition tracks, datasets, evaluation metrics, and the innovative approaches of the top-performing submissions, providing insights into the current advancements and future directions in the field of approximate nearest neighbor search.
  • PDF
    In thermalizing many-body quantum systems without conservation laws such as ergodic Floquet and random unitary circuits, local expectation values are predicted to decay to their equilibrium values exponentially quickly. In this work we derive a relationship between said exponential decay rate $\overline{g}$ and the operator spreading properties of a local unitary evolution. A hydrodynamical picture for operator spreading allows us to argue that, for random unitary circuits, $\overline{g}$ is encoded by the leading eigenvalue of a dynamical map obtained by enriching unitary dynamics with dissipation, in the limit of weak dissipation. We argue that the size of the eigenvalue does not depend on the details of this weak dissipation (given mild assumptions on properties of the ergodic dynamics), so long as it only suppresses large operators significantly. Our calculations are based on analytical results for random unitary circuits, but we argue that similar results hold for ergodic Floquet systems. These conjectures are in accordance with existing results which numerically obtain quantum many-body analogues of classical Ruelle-Pollicott resonances [T. Prosen J. Phys. A: Math. Gen. 35 L737 (2002), T. Mori, arXiv:2311.10304] by studying unitary evolutions subject to weak dissipation.
  • PDF
    Projective symmetries are ubiquitous in quantum lattice models and can be leveraged to constrain their phase diagram and entanglement structure. In this paper, we investigate the consequences of projective algebras formed by non-invertible symmetries and lattice translations in a generalized $1+1$D quantum XY model based on group-valued qudits. This model is specified by a finite group $G$ and enjoys a projective $\mathsf{Rep}(G)\times Z(G)$ and translation symmetry, where symmetry operators obey a projective algebra in the presence of symmetry defects. For invertible symmetries, such projective algebras imply Lieb-Schultz-Mattis (LSM) anomalies. However, this is not generally true for non-invertible symmetries, and we derive a condition on $G$ for the existence of an LSM anomaly. When this condition is not met, we prove that any unique and gapped ground state is necessarily a non-invertible weak symmetry protected topological (SPT) state with non-trivial entanglement, for which we construct an example fixed-point Hamiltonian. The projectivity also affects the dual symmetries after gauging $\mathsf{Rep}(G)\times Z(G)$ sub-symmetries, giving rise to non-Abelian and non-invertible dipole symmetries, as well as non-invertible translations. We complement our analysis with the SymTFT, where the projectivity causes it to be a topological order non-trivially enriched by translations. Throughout the paper, we develop techniques for gauging $\mathsf{Rep}(G)$ symmetry and inserting its symmetry defects on the lattice, which are applicable to other non-invertible symmetries.
  • PDF
    This paper describes a deep-SDM framework, MALPOLON. Written in Python and built upon the PyTorch library, this framework aims to facilitate training and inferences of deep species distribution models (deep-SDM) and sharing for users with only general Python language skills (e.g., modeling ecologists) who are interested in testing deep learning approaches to build new SDMs. More advanced users can also benefit from the framework's modularity to run more specific experiments by overriding existing classes while taking advantage of press-button examples to train neural networks on multiple classification tasks using custom or provided raw and pre-processed datasets. The framework is open-sourced on GitHub and PyPi along with extensive documentation and examples of use in various scenarios. MALPOLON offers straightforward installation, YAML-based configuration, parallel computing, multi-GPU utilization, baseline and foundational models for benchmarking, and extensive tutorials/documentation, aiming to enhance accessibility and performance scalability for ecologists and researchers.
  • PDF
    Self-supervised pretraining (SSP) has shown promising results in learning from large unlabeled datasets and, thus, could be useful for automated cardiovascular magnetic resonance (CMR) short-axis cine segmentation. However, inconsistent reports of the benefits of SSP for segmentation have made it difficult to apply SSP to CMR. Therefore, this study aimed to evaluate SSP methods for CMR cine segmentation. To this end, short-axis cine stacks of 296 subjects (90618 2D slices) were used for unlabeled pretraining with four SSP methods; SimCLR, positional contrastive learning, DINO, and masked image modeling (MIM). Subsets of varying numbers of subjects were used for supervised fine-tuning of 2D models for each SSP method, as well as to train a 2D baseline model from scratch. The fine-tuned models were compared to the baseline using the 3D Dice similarity coefficient (DSC) in a test dataset of 140 subjects. The SSP methods showed no performance gains with the largest supervised fine-tuning subset compared to the baseline (DSC = 0.89). When only 10 subjects (231 2D slices) are available for supervised training, SSP using MIM (DSC = 0.86) improves over training from scratch (DSC = 0.82). This study found that SSP is valuable for CMR cine segmentation when labeled training data is scarce, but does not aid state-of-the-art deep learning methods when ample labeled data is available. Moreover, the choice of SSP method is important. The code is publicly available at: https://github.com/q-cardIA/ssp-cmr-cine-segmentation
  • PDF
    We provide a rigorous analysis of the generalized Jang equation in the asymptotically anti-de Sitter setting modelled on constant time slices of anti-de Sitter spacetimes in dimensions $3\leq n \leq 7$ for a very general class of asymptotics. Potential applications to spacetime positive mass theorems for asymptotically anti-de Sitter initial data sets are discussed.
  • PDF
    This paper derives an achievable rate-distortion (R-D) region for the state-dependent discrete memoryless multiple access channel (SD-DMMAC), where the generalized feedback and causal side information are present at encoders, and the decoder performs the joint task of message decoding and state estimation. The Markov coding and backward-forward two-stage decoding schemes are adopted in the proof. This scenario is shown to be capable of modeling various integrated sensing and communication (ISAC) applications, including the monostatic-uplink system and multi-modal sensor networks, which are then studied as examples.
  • PDF
    Quantum Reservoir Computing (QRC) offers potential advantages over classical reservoir computing, including inherent processing of quantum inputs and a vast Hilbert space for state exploration. Yet, the relation between the performance of reservoirs based on complex and many-body quantum systems and non-classical state features is not established. Through an extensive analysis of QRC based on a transverse-field Ising model we show how different quantum effects, such as quantum coherence and correlations, contribute to improving the performance in temporal tasks, as measured by the Information Processing Capacity. Additionally, we critically assess the impact of finite measurement resources and noise on the reservoir's dynamics in different regimes, quantifying the limited ability to exploit quantum effects for increasing damping and noise strengths. Our results reveal a monotonic relationship between reservoir performance and coherence, along with the importance of quantum effects in the ergodic regime.
  • PDF
    We investigate a class of entanglement witnesses where each witness is formulated as a difference of two product observables. These observables are decomposable into positive semidefinite local operators that obey a partial ordering rule defined over all their possible expectation values. We provide a framework to construct these entanglement witnesses along with some examples. We also discuss methods to improve them both linearly and nonlinearly.
  • PDF
    Quantum simulators are ideal platforms to investigate quantum phenomena that are inaccessible through conventional means, such as the limited resources of classical computers to address large quantum systems or due to constraints imposed by fundamental laws of nature. Here, through a digitized adiabatic evolution, we report an experimental simulation of antiferromagnetic (AFM) and ferromagnetic (FM) phase formation induced by spontaneous symmetry breaking (SSB) in a three-generation Cayley tree-like superconducting lattice. We develop a digital quantum annealing algorithm to mimic the system dynamics, and observe the emergence of signatures of SSB-induced phase transition through a connected correlation function. We demonstrate that the signature of phase transition from classical AFM to quantum FM happens in systems undergoing zero-temperature adiabatic evolution with only nearest-neighbor interacting systems, the shortest range of interaction possible. By harnessing properties of the bipartite Renyi entropy as an entanglement witness, we observe the formation of entangled quantum FM and AFM phases. Our results open perspectives for new advances in condensed matter physics and digitized quantum annealing.
  • PDF
    Artificial Intelligence (AI), with its multiplier effect and wide applications in multiple areas, could potentially be an important application of quantum computing. Since modern AI systems are often built on neural networks, the design of quantum neural networks becomes a key challenge in integrating quantum computing into AI. To provide a more fine-grained characterisation of the impact of quantum components on the performance of neural networks, we propose a framework where classical neural network layers are gradually replaced by quantum layers that have the same type of input and output while keeping the flow of information between layers unchanged, different from most current research in quantum neural network, which favours an end-to-end quantum model. We start with a simple three-layer classical neural network without any normalisation layers or activation functions, and gradually change the classical layers to the corresponding quantum versions. We conduct numerical experiments on image classification datasets such as the MNIST, FashionMNIST and CIFAR-10 datasets to demonstrate the change of performance brought by the systematic introduction of quantum components. Through this framework, our research sheds new light on the design of future quantum neural network models where it could be more favourable to search for methods and frameworks that harness the advantages from both the classical and quantum worlds.
  • PDF
    Gravitational Rényi computations have traditionally been described in the language of Euclidean path integrals. In the semiclassical limit, such calculations are governed by Euclidean (or, more generally, complex) saddle-point geometries. We emphasize here that, at least in simple contexts, the Euclidean approach suggests an alternative formulation in terms of the bulk quantum wavefunction. Since this alternate formulation can be directly applied to the real-time quantum theory, it is insensitive to subtleties involved in defining the Euclidean path integral. In particular, it can be consistent with many different choices of integration contour. Despite the fact that self-adjoint operators in the associated real-time quantum theory have real eigenvalues, we note that the bulk wavefunction encodes the Euclidean (or complex) Rényi geometries that would arise in any Euclidean path integral. As a result, for any given quantum state, the appropriate real-time path integral yields both Rényi entropies and associated complex saddle-point geometries that agree with Euclidean methods. After brief explanations of these general points, we use JT gravity to illustrate the associated real-time computations in detail.
  • PDF
    $\mathbb{Z}_3$ lattice gauge theory is the simplest discrete gauge theory with three-quark bound states, i.e., baryons. Since it has a finite-dimensional Hilbert space, it can be used for testing quantum simulation of lattice gauge theory at nonzero baryon density. We discuss global and local gauge symmetries and their importance in quantum simulation. We perform quantum emulator calculation and demonstrate how to study the ground state property of baryonic matter.
  • PDF
    We propose a deterministic method to find all holographic entropy inequalities and prove the completeness of our method. We use a triality between holographic entropy inequalities, contraction maps and partial cubes. More specifically, the validity of a holographic entropy inequality is implied by the existence of a contraction map, which we prove to be equivalent to finding an isometric embedding of a contracted graph. Thus, by virtue of the completeness of the contraction map proof method, the problem of finding all holographic entropy inequalities is equivalent to the problem of finding all contraction maps, which we translate to a problem of finding all image graph partial cubes. We give an algorithmic solution to this problem and characterize the complexity of our method. We also demonstrate interesting by-products, most notably, a procedure to generate candidate quantum entropy inequalities.
  • PDF
    Proving the completeness of quantum mechanics has been a fundamental task since its foundation. After the formulation of the Bell inequalities, violated by quantum physics, it is nowadays believed that the theory is complete and non-local. While more general Bell-like inequalities, such as the one of Clauser and Horne, envisage a situation in which two parties choose at random two measurements to perform at causally-disconnected space-times, one could formulate temporal inequalities in which the two parties measure at different times. However, for causally-connected parties, these extensions are compatible with local hidden-variable theories, so that no quantum nature appears in such temporal correlations. Here we show that a temporal Clauser-Horne inequality for two spins is violated for nonzero time interval between the measurements if the two measured parties are connected by a spin chain. The chain constitutes a medium for the spreading of quantum information, which prevents the immediate signaling and thus the deterministic time evolution after the first measurement. Our result suggests that, as expected in a many-body setup, the Lieb-Robinson bound substitutes the speed of light as the fundamental limit for the spreading of information.
  • PDF
    We rigorously prove that in nearly arbitrary quantum spin chains with power-law-distributed random fields, namely such that the probability of a field exceeding $h$ scales as $h^{-\alpha}$, it is impossible for any operator evolving in the Heisenberg picture to spread with dynamical exponent less than $1/\alpha$. In particular, ballistic growth is impossible for $\alpha < 1$, diffusive growth is impossible for $\alpha < 1/2$, and any finite dynamical exponent becomes impossible for sufficiently small $\alpha$. This result thus establishes a wide family of models in which the disorder provably prevents conventional transport. We express the result as a tightening of Lieb-Robinson bounds due to random fields -- the proof modifies the standard derivation such that strong fields appear as effective weak interactions, and then makes use of analogous recent results for random-bond spin chains.
  • PDF
    For general random tensor network states at large bond dimension, we prove that the integer Rényi reflected entropies (away from phase transitions) are determined by minimal triway cuts through the network. This generalizes the minimal cut description of bipartite entanglement for these states. A natural extrapolation away from integer Rényi parameters, suggested by the triway cut problem, implies the holographic conjecture $S_R=2EW$, where $S_R$ is the reflected entropy and $EW$ is the entanglement wedge cross-section. Minimal triway cuts can be formulated as integer programs which cannot be relaxed to find a dual maximal flow/bit-thread description. This sheds light on the gap between the existence of tripartite entanglement in holographic states and the bipartite entanglement structure motivated by bit-threads. In particular, we prove that the Markov gap that measures tripartite entanglement is lower bounded by the integrality gap of the integer program that computes the triway cut.
  • PDF
    We investigate details of the chaotic dynamics of dual-unitary quantum circuits by evaluating all $2k$-point out-of-time-ordered correlators. For the generic class of circuits, by writing the correlators as contractions of a class of quantum channels, we prove their exponential decay. This implies that local operators separated in time approach free independence. Along the way, we develop a replica trick for dual-unitary circuits, which may be useful and of interest in its own right. We classify the relevant eigenstates of the replica transfer matrix by paths in the lattice of noncrossing partitions, combinatorial objects central to free probability theory. Interestingly, the noncrossing lattice emerges even for systems without randomness and with small onsite Hilbert space dimension.
  • PDF
    Understanding the physics of the two-dimensional Hubbard model is widely believed to be a key step in achieving a full understanding of high-$T_\mathrm{c}$ cuprate superconductors. In recent years, progress has been made by large-scale numerical simulations at finite doping and, on the other hand, by microscopic theories able to capture the physics of individual charge carriers. In this work, we study single pairs of dopants in a cylindrical system using the density-matrix renormalization group algorithm. We identify two coexisting charge configurations that couple to the spin environment in different ways: A tightly bound configuration featuring (next-)nearest-neighbor pairs and a stripe-like configuration of dopants on opposite sides of the cylinder, accompanied by a spin domain wall. Thus, we establish that the interplay between stripe order and uniform pairing, central to the models' phases at finite doping, has its origin at the single-pair level. By interpolating between the Hubbard and the related $t$-$J$ model, we are able to quantitatively understand discrepancies in the pairing properties of the two models through the three-site hopping term usually omitted from the $t$-$J$ Hamiltonian. This term is closely related to a next-nearest-neighbor tunneling $t'$, which we observe to upset the balance between the competing stripe and pair states on the two-dopant level.
  • PDF
    The Neptunian desert and savanna have been recently found to be separated by a ridge, an overdensity of planets in the $\simeq$3-5 days period range. These features are thought to be shaped by dynamical and atmospheric processes. However, their relative roles are not yet well understood. We intend to confirm and characterise the super-Neptune TESS candidate TOI-5005.01, which orbits a moderately bright (V = 11.8) solar-type star (G2 V) with an orbital period of 6.3 days. We confirm TOI-5005 b to be a transiting super-Neptune with a radius of $R_{\rm p}$ = $6.25\pm 0.24$ $\rm R_{\rm \oplus}$ ($R_{\rm p}$ = $0.558\pm 0.021$ $\rm R_{\rm J}$) and a mass of $M_{\rm p}$ = $32.7\pm 5.9$ $\rm M_{\oplus}$ ($M_{\rm p}$ = $0.103\pm 0.018$ $\rm M_{\rm J}$), which corresponds to a mean density of $\rho_{\rm p}$ = $0.74 \pm 0.16$ $\rm g \, cm^{-3}$. Our internal structure modelling indicates that the overall metal mass fraction is well constrained to a value slightly lower than that of Neptune and Uranus ($Z_{\rm planet}$ = $0.76^{+0.04}_{-0.11}$). We also estimated the present-day atmospheric mass-loss rate of TOI-5005 b but found contrasting predictions depending on the choice of photoevaporation model. At a population level, we find statistical evidence ($p$-value = $0.0092^{+0.0184}_{-0.0066}$) that planets in the savanna such as TOI-5005 b tend to show lower densities than planets in the ridge, with a dividing line around 1 $\rm g \, cm^{-3}$, which supports the hypothesis of different evolutionary pathways populating both regimes. TOI-5005 b is located in a key region of the period-radius space to study the transition between the Neptunian ridge and the savanna. It orbits the brightest star of all such planets, which makes it a target of interest for atmospheric and orbital architecture observations that will bring a clearer picture of its overall evolution.
  • PDF
    The high-temperature limit of the superconformal index, especially on higher sheets, often captures useful universal information about a theory. In 4d $\mathcal{N}=2$ superconformal field theories with fractional r-charges, there exists a special notion of high-temperature limit on higher sheets that captures data of three-dimensional topological quantum field theories arising from r-twisted circle reduction. These TQFTs are closely tied with the VOA of the 4d SCFT. We study such high-temperature limits. More specifically, we apply Di~Pietro-Komargodski type supersymmetric effective field theory techniques to r-twisted circle reductions of $(A_1,A_{2n})$ Argyres-Douglas theories, leveraging their Maruyoshi-Song Lagrangian with manifest $\mathcal{N}=1$ supersymmetry. The result on the second sheet is the Gang-Kim-Stubbs family of 3d $\mathcal{N}=2$ SUSY enhancing rank-$0$ theories with monopole superpotentials, whose boundary supports the Virasoro minimal model VOAs $M(2,2n+3)$. Upon topological twist, they give non-unitary TQFTs controlled by the $M(2,2n+3)$ modular tensor category (MTC). The high-temperature limit on other sheets yields their unitary or non-unitary Galois conjugates. This opens up the prospect of a broader four-supercharge perspective on the celebrated correspondence between 4d $\mathcal{N}=2$ SCFTs and 2d VOAs via interpolating 3d EFTs. Several byproducts follow, including a systematic approach to 3d SUSY enhancement from 4d SUSY enhancement, and a 3d QFT handle on Galois orbits of various MTCs associated with 4d $\mathcal{N}=2$ SCFTs.
  • PDF
    Building on the success of diffusion models in visual generation, flow-based models reemerge as another prominent family of generative models that have achieved competitive or better performance in terms of both visual quality and inference speed. By learning the velocity field through flow-matching, flow-based models tend to produce a straighter sampling trajectory, which is advantageous during the sampling process. However, unlike diffusion models for which fast samplers are well-developed, efficient sampling of flow-based generative models has been rarely explored. In this paper, we propose a framework called FlowTurbo to accelerate the sampling of flow-based models while still enhancing the sampling quality. Our primary observation is that the velocity predictor's outputs in the flow-based models will become stable during the sampling, enabling the estimation of velocity via a lightweight velocity refiner. Additionally, we introduce several techniques including a pseudo corrector and sample-aware compilation to further reduce inference time. Since FlowTurbo does not change the multi-step sampling paradigm, it can be effectively applied for various tasks such as image editing, inpainting, etc. By integrating FlowTurbo into different flow-based models, we obtain an acceleration ratio of 53.1%$\sim$58.3% on class-conditional generation and 29.8%$\sim$38.5% on text-to-image generation. Notably, FlowTurbo reaches an FID of 2.12 on ImageNet with 100 (ms / img) and FID of 3.93 with 38 (ms / img), achieving the real-time image generation and establishing the new state-of-the-art. Code is available at https://github.com/shiml20/FlowTurbo.
  • PDF
    As the prevalence of wearable devices, learning egocentric motions becomes essential to develop contextual AI. In this work, we present EgoLM, a versatile framework that tracks and understands egocentric motions from multi-modal inputs, e.g., egocentric videos and motion sensors. EgoLM exploits rich contexts for the disambiguation of egomotion tracking and understanding, which are ill-posed under single modality conditions. To facilitate the versatile and multi-modal framework, our key insight is to model the joint distribution of egocentric motions and natural languages using large language models (LLM). Multi-modal sensor inputs are encoded and projected to the joint latent space of language models, and used to prompt motion generation or text generation for egomotion tracking or understanding, respectively. Extensive experiments on large-scale multi-modal human motion dataset validate the effectiveness of EgoLM as a generalist model for universal egocentric learning.
  • PDF
    It has been proposed that diabatic quantum annealing (DQA), which turns off the transverse field at a finite speed, produces samples well described by the Boltzmann distribution. We analytically show that, up to linear order in quenching time, the DQA approximates a high-temperature Boltzmann distribution. Our theory yields an explicit relation between the quenching rate of the DQA and the temperature of the Boltzmann distribution. Based on this result, we discuss how the DQA can be utilized to train the Restricted Boltzmann Machine (RBM).
  • PDF
    Recent advancements in Large Multimodal Models (LMMs) have greatly enhanced their proficiency in 2D visual understanding tasks, enabling them to effectively process and understand images and videos. However, the development of LMMs with 3D-awareness for 3D scene understanding has been hindered by the lack of large-scale 3D vision-language datasets and powerful 3D encoders. In this paper, we introduce a simple yet effective framework called LLaVA-3D. Leveraging the strong 2D understanding priors from LLaVA, our LLaVA-3D efficiently adapts LLaVA for 3D scene understanding without compromising 2D understanding capabilities. To achieve this, we employ a simple yet effective representation, 3D Patch, which connects 2D CLIP patch features with their corresponding positions in 3D space. By integrating the 3D Patches into 2D LMMs and employing joint 2D and 3D vision-language instruction tuning, we establish a unified architecture for both 2D image understanding and 3D scene understanding. Experimental results show that LLaVA-3D converges 3.5x faster than existing 3D LMMs when trained on 3D vision-language datasets. Moreover, LLaVA-3D not only achieves state-of-the-art performance across various 3D tasks but also maintains comparable 2D image understanding and vision-language conversation capabilities with LLaVA.
  • PDF
    Leveraging the visual priors of pre-trained text-to-image diffusion models offers a promising solution to enhance zero-shot generalization in dense prediction tasks. However, existing methods often uncritically use the original diffusion formulation, which may not be optimal due to the fundamental differences between dense prediction and image generation. In this paper, we provide a systemic analysis of the diffusion formulation for the dense prediction, focusing on both quality and efficiency. And we find that the original parameterization type for image generation, which learns to predict noise, is harmful for dense prediction; the multi-step noising/denoising diffusion process is also unnecessary and challenging to optimize. Based on these insights, we introduce Lotus, a diffusion-based visual foundation model with a simple yet effective adaptation protocol for dense prediction. Specifically, Lotus is trained to directly predict annotations instead of noise, thereby avoiding harmful variance. We also reformulate the diffusion process into a single-step procedure, simplifying optimization and significantly boosting inference speed. Additionally, we introduce a novel tuning strategy called detail preserver, which achieves more accurate and fine-grained predictions. Without scaling up the training data or model capacity, Lotus achieves SoTA performance in zero-shot depth and normal estimation across various datasets. It also significantly enhances efficiency, being hundreds of times faster than most existing diffusion-based methods.
  • PDF
    We establish property $R_\infty$ for Artin groups of spherical type $D_n$, $n\ge6$, their central quotients, and also for large hyperbolic-type free-of-infinity Artin groups and some other classes of large-type Artin groups. The key ingredients are recent descriptions of the automorphism groups for these Artin groups and their action on suitable Gromov-hyperbolic spaces.
  • PDF
    We propose a framework for active mapping and exploration that leverages Gaussian splatting for constructing information-rich maps. Further, we develop a parallelized motion planning algorithm that can exploit the Gaussian map for real-time navigation. The Gaussian map constructed onboard the robot is optimized for both photometric and geometric quality while enabling real-time situational awareness for autonomy. We show through simulation experiments that our method is competitive with approaches that use alternate information gain metrics, while being orders of magnitude faster to compute. In real-world experiments, our algorithm achieves better map quality (10% higher Peak Signal-to-Noise Ratio (PSNR) and 30% higher geometric reconstruction accuracy) than Gaussian maps constructed by traditional exploration baselines. Experiment videos and more details can be found on our project page: https://tyuezhan.github.io/RT_GuIDE/
  • PDF
    Humans can learn to manipulate new objects by simply watching others; providing robots with the ability to learn from such demonstrations would enable a natural interface specifying new behaviors. This work develops Robot See Robot Do (RSRD), a method for imitating articulated object manipulation from a single monocular RGB human demonstration given a single static multi-view object scan. We first propose 4D Differentiable Part Models (4D-DPM), a method for recovering 3D part motion from a monocular video with differentiable rendering. This analysis-by-synthesis approach uses part-centric feature fields in an iterative optimization which enables the use of geometric regularizers to recover 3D motions from only a single video. Given this 4D reconstruction, the robot replicates object trajectories by planning bimanual arm motions that induce the demonstrated object part motion. By representing demonstrations as part-centric trajectories, RSRD focuses on replicating the demonstration's intended behavior while considering the robot's own morphological limits, rather than attempting to reproduce the hand's motion. We evaluate 4D-DPM's 3D tracking accuracy on ground truth annotated 3D part trajectories and RSRD's physical execution performance on 9 objects across 10 trials each on a bimanual YuMi robot. Each phase of RSRD achieves an average of 87% success rate, for a total end-to-end success rate of 60% across 90 trials. Notably, this is accomplished using only feature fields distilled from large pretrained vision models -- without any task-specific training, fine-tuning, dataset collection, or annotation. Project page: https://robot-see-robot-do.github.io
  • PDF
    Traditionally, unmanned aerial vehicles (UAVs) rely on CMOS-based cameras to collect images about the world below. One of the most successful applications of UAVs is to generate orthomosaics or orthomaps, in which a series of images are integrated together to develop a larger map. However, the use of CMOS-based cameras with global or rolling shutters mean that orthomaps are vulnerable to challenging light conditions, motion blur, and high-speed motion of independently moving objects under the camera. Event cameras are less sensitive to these issues, as their pixels are able to trigger asynchronously on brightness changes. This work introduces the first orthomosaic approach using event cameras. In contrast to existing methods relying only on CMOS cameras, our approach enables map generation even in challenging light conditions, including direct sunlight and after sunset.
  • PDF
    Contrastive Language-Image Pre-training (CLIP) shows promise in medical image analysis but requires substantial data and computational resources. Due to these restrictions, existing CLIP applications in medical imaging focus mainly on modalities like chest X-rays that have abundant image-report data available, leaving many other important modalities under-explored. Here, we propose the first adaptation of the full CLIP model to mammography, which presents significant challenges due to labeled data scarcity, high-resolution images with small regions of interest, and data imbalance. We first develop a specialized supervision framework for mammography that leverages its multi-view nature. Furthermore, we design a symmetric local alignment module to better focus on detailed features in high-resolution images. Lastly, we incorporate a parameter-efficient fine-tuning approach for large language models pre-trained with medical knowledge to address data limitations. Our multi-view and multi-scale alignment (MaMA) method outperforms state-of-the-art baselines for three different tasks on two large real-world mammography datasets, EMBED and RSNA-Mammo, with only 52% model size compared with the largest baseline.
  • PDF
    We develop formal privacy mechanisms for releasing statistics from data with many outlying values, such as income data. These mechanisms ensure that a per-record differential privacy guarantee degrades slowly in the protected records' influence on the statistics being released. Formal privacy mechanisms generally add randomness, or "noise," to published statistics. If a noisy statistic's distribution changes little with the addition or deletion of a single record in the underlying dataset, an attacker looking at this statistic will find it plausible that any particular record was present or absent, preserving the records' privacy. More influential records -- those whose addition or deletion would change the statistics' distribution more -- typically suffer greater privacy loss. The per-record differential privacy framework quantifies these record-specific privacy guarantees, but existing mechanisms let these guarantees degrade rapidly (linearly or quadratically) with influence. While this may be acceptable in cases with some moderately influential records, it results in unacceptably high privacy losses when records' influence varies widely, as is common in economic data. We develop mechanisms with privacy guarantees that instead degrade as slowly as logarithmically with influence. These mechanisms allow for the accurate, unbiased release of statistics, while providing meaningful protection for highly influential records. As an example, we consider the private release of sums of unbounded establishment data such as payroll, where our mechanisms extend meaningful privacy protection even to very large establishments. We evaluate these mechanisms empirically and demonstrate their utility.
  • PDF
    Proxy pattern-mixture models (PPMM) have previously been proposed as a model-based framework for assessing the potential for nonignorable nonresponse in sample surveys and nonignorable selection in nonprobability samples. One defining feature of the PPMM is the single sensitivity parameter, $\phi$, that ranges from 0 to 1 and governs the degree of departure from ignorability. While this sensitivity parameter is attractive in its simplicity, it may also be of interest to describe departures from ignorability in terms of how the odds of response (or selection) depend on the outcome being measured. In this paper, we re-express the PPMM as a selection model, using the known relationship between pattern-mixture models and selection models, in order to better understand the underlying assumptions of the PPMM and the implied effect of the outcome on nonresponse. The selection model that corresponds to the PPMM is a quadratic function of the survey outcome and proxy variable, and the magnitude of the effect depends on the value of the sensitivity parameter, $\phi$ (missingness/selection mechanism), the differences in the proxy means and standard deviations for the respondent and nonrespondent populations, and the strength of the proxy, $\rho^{(1)}$. Large values of $\phi$ (beyond $0.5$) often result in unrealistic selection mechanisms, and the corresponding selection model can be used to establish more realistic bounds on $\phi$. We illustrate the results using data from the U.S. Census Household Pulse Survey.

Recent comments

Chia-Ying Lin Sep 27 2024 18:47 UTC

Hi! I found this paper is cool, and would like to know where to find the proof of the lemma 3.7 and 3.8?

Angelo Lucia Sep 21 2024 12:30 UTC

Aram is correct: we roughly prove that if you can show a slower than 1/n^2 lower bound to the gap, you can bootstrap it to a constant bound. But if the gap closes faster than you don't get any improvement.

Aram Harrow Sep 21 2024 12:02 UTC

The gap can vanish faster than 1/n^2. Their theorem just says it can’t vanish more slowly. See eq 8.

Varun Menon Sep 20 2024 14:03 UTC

Nice result! I am wondering how to reconcile this result with this example of the 'area weighted Motzkin chains' https://arxiv.org/abs/1611.03147 where the model is frustration free, the Hamiltonian is a sum of projectors, and has short range interactions everywhere (even at the boundaries where the

...(continued)
Unme Sep 19 2024 14:17 UTC

Thanks for the interesting paper!

I would like to ask whether the condition in Theorem 6 should be $PE_i^{\dagger}E_jP=\alpha_{ij}P$?

Mark M. Wilde Sep 19 2024 06:23 UTC

Thanks for your reply and for your comment.

Indeed it is true that the mentioned papers do not use the phrase "superchannel", which could lead to one missing these references. They instead use the phrase "bipartite operation", which is a less fitting term.

D.-S. Wang Sep 19 2024 05:58 UTC

Dear Mark, thank you for letting us know the papers and sorry for replying late. I managed to create an account. We will definitely cite them in future work if they are relevant on a technical level. My arXiv posting is random. Given my limited knowledge, probably I wouldn't encounter them in the co

...(continued)
Mark M. Wilde Sep 18 2024 11:27 UTC

This kind of theory, thinking of codes in terms of superchannels, was developed previously in https://arxiv.org/abs/1406.7142 , which is not cited in the paper. See also the related work https://arxiv.org/abs/1210.4722 . The paper is now published before the arxiv posting appeared, which is unfortun

...(continued)
MariusK Sep 16 2024 14:30 UTC

Your *Better Solution Probability* (BSP) metric reminds me of the concept of *advantage* used in reinforcement learning. There, given a *reward* (or better, a *discounted return*) $R$, the advantage is defined via $\mathrm{adv} = R - \mathrm{base}$. Here, $\mathrm{base}$ is a baseline. It could be a

...(continued)
John Martyn Sep 11 2024 19:05 UTC

Hi Seiseki, thanks for the comment! Indeed, as our desired map is not trace-preserving, we prove Lemma 2 in order to apply randomized compiling to QSP. This offers a quadratic suppression of error in general.

As you mention, your lemmas 4.1 and 4.2 upper bound the error of a randomly-compiled chan

...(continued)