22 research outputs found
Scaled Particle Theory for Hard Sphere Pairs. II. Numerical Analysis
We use the extension of scaled particle theory (ESPT) presented in the
accompanying paper [Stillinger et al. J. Chem. Phys. xxx, xxx (2007)] to
calculate numerically pair correlation function of the hard sphere fluid over
the density range . Comparison with computer
simulation results reveals that the new theory is able to capture accurately
the fluid's structure across the entire density range examined. The pressure
predicted via the virial route is systematically lower than simulation results,
while that obtained using the compressibility route is lower than simulation
predictions for and higher than simulation predictions
for . Numerical predictions are also presented for the
surface tension and Tolman length of the hard sphere fluid
Q-Pandora Unboxed: Characterizing Noise Resilience of Quantum Error Correction Codes
Quantum error correction codes (QECCs) are critical for realizing reliable
quantum computing by protecting fragile quantum states against noise and
errors. However, limited research has analyzed the noise resilience of QECCs to
help select optimal codes. This paper conducts a comprehensive study analyzing
two QECCs - rotated and unrotated surface codes - under different error types
and noise models using simulations. Among them, rotated surface codes perform
best with higher thresholds attributed to simplicity and lower qubit overhead.
The noise threshold, or the point at which QECCs become ineffective, surpasses
the error rate found in contemporary quantum processors. When confronting
quantum hardware where a specific error or noise model is dominant, a
discernible hierarchy emerges for surface code implementation in terms of
resource demand. This ordering is consistently observed across unrotated, and
rotated surface codes. Our noise model analysis ranks the code-capacity model
as the most pessimistic and circuit-level model as the most realistic. The
study maps error thresholds, revealing surface code's advantage over modern
quantum processors. It also shows higher code distances and rounds consistently
improve performance. However, excessive distances needlessly increase qubit
overhead. By matching target logical error rates and feasible number of qubits
to optimal surface code parameters, our study demonstrates the necessity of
tailoring these codes to balance reliability and qubit resources. Conclusively,
we underscore the significance of addressing the notable challenges associated
with surface code overheads and qubit improvements.Comment: 15 pages; 9 figures; 3 table
Quantum Error Correction For Dummies
In the current Noisy Intermediate Scale Quantum (NISQ) era of quantum
computing, qubit technologies are prone to imperfections, giving rise to
various errors such as gate errors, decoherence/dephasing, measurement errors,
leakage, and crosstalk. These errors present challenges in achieving error-free
computation within NISQ devices. A proposed solution to this issue is Quantum
Error Correction (QEC), which aims to rectify the corrupted qubit state through
a three-step process: (i) detection: identifying the presence of an error, (ii)
decoding: pinpointing the location(s) of the affected qubit(s), and (iii)
correction: restoring the faulty qubits to their original states. QEC is an
expanding field of research that encompasses intricate concepts. In this paper,
we aim to provide a comprehensive review of the historical context, current
state, and future prospects of Quantum Error Correction, tailored to cater to
computer scientists with limited familiarity with quantum physics and its
associated mathematical concepts. In this work, we, (a) explain the
foundational principles of QEC and explore existing Quantum Error Correction
Codes (QECC) designed to correct errors in qubits, (b) explore the practicality
of these QECCs concerning implementation and error correction quality, and (c)
highlight the challenges associated with implementing QEC within the context of
the current landscape of NISQ computers.Comment: 12 pages, 9 figures, 4 table
Quantum Random Access Memory For Dummies
Quantum Random Access Memory (QRAM) has the potential to revolutionize the
area of quantum computing. QRAM uses quantum computing principles to store and
modify quantum or classical data efficiently, greatly accelerating a wide range
of computer processes. Despite its importance, there is a lack of comprehensive
surveys that cover the entire spectrum of QRAM architectures. We fill this gap
by providing a comprehensive review of QRAM, emphasizing its significance and
viability in existing noisy quantum computers. By drawing comparisons with
conventional RAM for ease of understanding, this survey clarifies the
fundamental ideas and actions of QRAM.Comment: 12 pages, 10 figures, 4 tables, 65 citation
Scaled Particle Theory for Hard Sphere Pairs. I. Mathematical Structure
We develop an extension of the original Reiss-Frisch-Lebowitz scaled particle
theory that can serve as a predictive method for the hard sphere pair
correlation function g(r). The reversible cavity creation work is analyzed both
for a single spherical cavity of arbitrary size, as well as for a pair of
identical such spherical cavities with variable center-to-center separation.
These quantities lead directly to prediction of g(r). Smooth connection
conditions have been identified between the small-cavity situation where the
work can be exactly and completely expressed in terms of g(r), and the
large-cavity regime where macroscopic properties become relevant. Closure
conditions emerge which produce a nonlinear integral equation that must be
satisfied by the pair correlation function. This integral equation has a
structure which straightforwardly generates a solution that is a power series
in density. The results of this series replicate the exact second and third
virial coefficients for the hard sphere system via the contact value of the
pair correlation function. The predicted fourth virial coefficient is
approximately 0.6 percent lower than the known exact value. Detailed numerical
analysis of the nonlinear integral equation has been deferred to the sequel
(following paper
Deep phenotyping and genomic data from a nationally representative study on dementia in India
The Harmonized Diagnostic Assessment of Dementia for the Longitudinal Aging Study in India (LASI-DAD) is a nationally representative in-depth study of cognitive aging and dementia. We present a publicly available dataset of harmonized cognitive measures of 4,096 adults 60 years of age and older in India, collected across 18 states and union territories. Blood samples were obtained to carry out whole blood and serum-based assays. Results are included in a venous blood specimen datafile that can be linked to the Harmonized LASI-DAD dataset. A global screening array of 960 LASI-DAD respondents is also publicly available for download, in addition to neuroimaging data on 137 LASI-DAD participants. Altogether, these datasets provide comprehensive information on older adults in India that allow researchers to further understand risk factors associated with cognitive impairment and dementia.Peer reviewe