468 research outputs found
Sampling functions for multimode homodyne tomography with a single local oscillator
We derive various sampling functions for multimode homodyne tomography with a
single local oscillator. These functions allow us to sample multimode
s-parametrized quasidistributions, density matrix elements in Fock basis, and
s-ordered moments of arbitrary order directly from the measured quadrature
statistics. The inevitable experimental losses can be compensated by proper
modification of the sampling functions. Results of Monte Carlo simulations for
squeezed three-mode state are reported and the feasibility of reconstruction of
the three-mode Q-function and s-ordered moments from 10^7 sampled data is
demonstrated.Comment: 12 pages, 8 figures, REVTeX, submitted Phys. Rev.
Conditional large Fock state preparation and field state reconstruction in Cavity QED
We propose a scheme for producing large Fock states in Cavity QED via the
implementation of a highly selective atom-field interaction. It is based on
Raman excitation of a three-level atom by a classical field and a quantized
field mode. Selectivity appears when one tunes to resonance a specific
transition inside a chosen atom-field subspace, while other transitions remain
dispersive, as a consequence of the field dependent electronic energy shifts.
We show that this scheme can be also employed for reconstructing, in a new and
efficient way, the Wigner function of the cavity field state.Comment: 4 Revtex pages with 3 postscript figures. Submitted for publicatio
Quantum inference of states and processes
The maximum-likelihood principle unifies inference of quantum states and
processes from experimental noisy data. Particularly, a generic quantum process
may be estimated simultaneously with unknown quantum probe states provided that
measurements on probe and transformed probe states are available. Drawbacks of
various approximate treatments are considered.Comment: 7 pages, 4 figure
Phase-space formulation of quantum mechanics and quantum state reconstruction for physical systems with Lie-group symmetries
We present a detailed discussion of a general theory of phase-space
distributions, introduced recently by the authors [J. Phys. A {\bf 31}, L9
(1998)]. This theory provides a unified phase-space formulation of quantum
mechanics for physical systems possessing Lie-group symmetries. The concept of
generalized coherent states and the method of harmonic analysis are used to
construct explicitly a family of phase-space functions which are postulated to
satisfy the Stratonovich-Weyl correspondence with a generalized traciality
condition. The symbol calculus for the phase-space functions is given by means
of the generalized twisted product. The phase-space formalism is used to study
the problem of the reconstruction of quantum states. In particular, we consider
the reconstruction method based on measurements of displaced projectors, which
comprises a number of recently proposed quantum-optical schemes and is also
related to the standard methods of signal processing. A general group-theoretic
description of this method is developed using the technique of harmonic
expansions on the phase space.Comment: REVTeX, 18 pages, no figure
Euclid Preparation TBD. Characterization of convolutional neural networks for the identification of galaxy-galaxy strong lensing events
Forthcoming imaging surveys will increase the number of known galaxy-scale strong lenses by several orders of magnitude. For this to happen, images of billions of galaxies will have to be inspected to identify potential candidates. In this context, deep-learning techniques are particularly suitable for finding patterns in large data sets, and convolutional neural networks (CNNs) in particular can efficiently process large volumes of images. We assess and compare the performance of three network architectures in the classification of strong-lensing systems on the basis of their morphological characteristics. In particular, we implemented a classical CNN architecture, an inception network, and a residual network. We trained and tested our networks on different subsamples of a data set of 40 000 mock images whose characteristics were similar to those expected in the wide survey planned with the ESA mission Euclid, gradually including larger fractions of faint lenses. We also evaluated the importance of adding information about the color difference between the lens and source galaxies by repeating the same training on single- and multiband images. Our models find samples of clear lenses with âł90% precision and completeness. Nevertheless, when lenses with fainter arcs are included in the training set, the performance of the three models deteriorates with accuracy values of ~0.87 to ~0.75, depending on the model. Specifically, the classical CNN and the inception network perform similarly in most of our tests, while the residual network generally produces worse results. Our analysis focuses on the application of CNNs to high-resolution space-like images, such as those that the Euclid telescope will deliver. Moreover, we investigated the optimal training strategy for this specific survey to fully exploit the scientific potential of the upcoming observations. We suggest that training the networks separately on lenses with different morphology might be needed to identify the faint arcs. We also tested the relevance of the color information for the detection of these systems, and we find that it does not yield a significant improvement. The accuracy ranges from ~0.89 to ~0.78 for the different models. The reason might be that the resolution of the Euclid telescope in the infrared bands is lower than that of the images in the visual band
Euclid preparation. XXIV. Calibration of the halo mass function in Î(Îœ)CDM cosmologies
Euclidâs photometric galaxy cluster survey has the potential to be a very competitive cosmological probe. The main cosmological probe with observations of clusters is their number count, within which the halo mass function (HMF) is a key theoretical quantity. We present a new calibration of the analytic HMF, at the level of accuracy and precision required for the uncertainty in this quantity to be subdominant with respect to other sources of uncertainty in recovering cosmological parameters from Euclid cluster counts. Our model is calibrated against a suite of N-body simulations using a Bayesian approach taking into account systematic errors arising from numerical effects in the simulation. First, we test the convergence of HMF predictions from different N-body codes, by using initial conditions generated with different orders of Lagrangian Perturbation theory, and adopting different simulation box sizes and mass resolution. Then, we quantify the effect of using different halo finder algorithms, and how the resulting differences propagate to the cosmological constraints. In order to trace the violation of universality in the HMF, we also analyse simulations based on initial conditions characterised by scale-free power spectra with different spectral indexes, assuming both Einsteinâde Sitter and standard ÎCDM expansion histories. Based on these results, we construct a fitting function for the HMF that we demonstrate to be sub-percent accurate in reproducing results from 9 different variants of the ÎCDM model including massive neutrinos cosmologies. The calibration systematic uncertainty is largely sub-dominant with respect to the expected precision of future massâobservation relations; with the only notable exception of the effect due to the halo finder, that could lead to biased cosmological inference
Quantum state reconstruction using atom optics
We present a novel technique in which the total internal quantum state of an atom may be reconstructed via the measurement of the momentum transferred to an atom following its interaction with a near resonant travelling wave laser beam. We present the first such measurement and demonstrate the feasibility of the technique
Euclid: modelling massive neutrinos in cosmology - a code comparison
Material outgassing in a vacuum leads to molecular contamination, a well-known problem in spaceflight. Water is the most common contaminant in cryogenic spacecraft, altering numerous properties of optical systems. Too much ice means that Euclidâs calibration requirements cannot be met anymore. Euclid must then be thermally decontaminated, which is a month-long risky operation. We need to understand how ice affects our data to build adequate calibration and survey plans. A comprehensive analysis in the context of an astrophysical space survey has not been done before. In this paper we look at other spacecraft with well-documented outgassing records. We then review the formation of thin ice films, and find that for Euclid a mix of amorphous and crystalline ices is expected. Their surface topography â and thus optical properties â depend on the competing energetic needs of the substrate-water and the water-water interfaces, and they are hard to predict with current theories. We illustrate that with scanning-tunnelling and atomic-force microscope images of thin ice films. Sophisticated tools exist to compute contamination rates, and we must understand their underlying physical principles and uncertainties. We find considerable knowledge errors on the diffusion and sublimation coefficients, limiting the accuracy of outgassing estimates. We developed a water transport model to compute contamination rates in Euclid, and find agreement with industry estimates within the uncertainties. Tests of the Euclid flight hardware in space simulators did not pick up significant contamination signals, but they were also not geared towards this purpose; our in-flight calibration observations will be much more sensitive. To derive a calibration and decontamination strategy, we need to understand the link between the amount of ice in the optics and its effect on the data. There is little research about this, possibly because other spacecraft can decontaminate more easily, quenching the need for a deeper understanding. In our second paper, we quantify the impact of iced optics on Euclidâs data
Euclid preparation:XXIII. Derivation of galaxy physical properties with deep machine learning using mock fluxes and H-band images
Next generation telescopes, such as Euclid, Rubin/LSST, and Roman, will open new windows on the Universe, allowing us to infer physical properties for tens of millions of galaxies. Machine learning methods are increasingly becoming the most efficient tools to handle this enormous amount of data, not only as they are faster to apply to data samples than traditional methods, but because they are also often more accurate. Properly understanding their applications and limitations for the exploitation of these data is of utmost importance. In this paper we present an exploration of this topic by investigating how well redshifts, stellar masses, and star-formation rates can be measured with deep learning algorithms for galaxies within data that mimics the Euclid and Rubin/LSST surveys. We find that Deep Learning Neural Networks and Convolutional Neutral Networks (CNN), which are dependent on the parameter space of the sample used for training, perform well in measuring the properties of these galaxies and have an accuracy which is better than traditional methods based on spectral energy distribution fitting. CNNs allow the processing of multi-band magnitudes together with -band images. We find that the estimates of stellar masses improve with the use of an image, but those of redshift and star-formation rates do not. Our best machine learning results are deriving i) the redshift within a normalised error of less than 0.15 for 99.9% of the galaxies in the sample with S/N>3 in the -band; ii) the stellar mass within a factor of two (0.3 dex) for 99.5% of the considered galaxies; iii) the star-formation rates within a factor of two (0.3 dex) for 70% of the sample. We discuss the implications of our work for application to surveys, mainly but not limited to Euclid and Rubin/LSST, and how measurements of these galaxy parameters can be improved with deep learning
Euclid Preparation. TBD. Impact of magnification on spectroscopic galaxy clustering
In this paper we investigate the impact of lensing magnification on the analysis of Euclid's spectroscopic survey, using the multipoles of the 2-point correlation function for galaxy clustering. We determine the impact of lensing magnification on cosmological constraints, and the expected shift in the best-fit parameters if magnification is ignored. We consider two cosmological analyses: i) a full-shape analysis based on the CDM model and its extension CDM and ii) a model-independent analysis that measures the growth rate of structure in each redshift bin. We adopt two complementary approaches in our forecast: the Fisher matrix formalism and the Markov chain Monte Carlo method. The fiducial values of the local count slope (or magnification bias), which regulates the amplitude of the lensing magnification, have been estimated from the Euclid Flagship simulations. We use linear perturbation theory and model the 2-point correlation function with the public code coffe. For a CDM model, we find that the estimation of cosmological parameters is biased at the level of 0.4-0.7 standard deviations, while for a CDM dynamical dark energy model, lensing magnification has a somewhat smaller impact, with shifts below 0.5 standard deviations. In a model-independent analysis aiming to measure the growth rate of structure, we find that the estimation of the growth rate is biased by up to standard deviations in the highest redshift bin. As a result, lensing magnification cannot be neglected in the spectroscopic survey, especially if we want to determine the growth factor, one of the most promising ways to test general relativity with Euclid. We also find that, by including lensing magnification with a simple template, this shift can be almost entirely eliminated with minimal computational overhead
- âŠ