467 research outputs found

    Astrophotonic micro-spectrographs in the era of ELTs

    Full text link
    The next generation of Extremely Large Telescopes (ELT), with diameters up to 39 meters, will start opera- tion in the next decade and promises new challenges in the development of instruments. The growing field of astrophotonics (the use of photonic technologies in astronomy) can partly solve this problem by allowing mass production of fully integrated and robust instruments combining various optical functions, with the potential to reduce the size, complexity and cost of instruments. In this paper, we focus on developments in integrated micro-spectrographs and their potential for ELTs. We take an inventory of the identified technologies currently in development, and compare the performance of the different concepts. We show that in the current context of single-mode instruments, integrated spectrographs making use of, e.g., a photonic lantern can be a solution to reach the desired performance. However, in the longer term, there is a clear need to develop multimode devices to improve overall the throughput and sensitivity, while decreasing the instrument complexity.Comment: 9 pages. 2 figures. Proceeding of SPIE 9147 "Ground-based and Airborne Instrumentation for Astronomy V

    Evaluating the demand side: New challenges for evaluation

    Get PDF
    Evaluation of research and innovation policy faces radical challenges arising from a new policy emphasis upon demand-side measures and linked to this an understanding of innovation policy as a means to achieve societal goals. This article considers the implications for the practice of evaluation at both micro and meso-levels. It uses the exemplar of an evaluation design for the European Union's Lead Market Initiative to expose the extent to which classical approaches to evaluation are valid and where new issues arise. Some problems highlighted include the difficulty of establishing a relevant baseline, the inability of public statistics constructed in supply-side mode to capture actions, the need to engage with actors who do not necessarily see themselves as part of the initiative being evaluated, long timescales and potential wide geographical scope, measures that span from micro to macro, and blurred boundaries between implementation and impact. It is concluded that there is a key role for evaluators to become involved in co-learning and co-evolution of these policy instruments in a manner analogous to the relationship between evaluation and policy development that characterized the emergence of collaborative R&D support programmes

    Reconstructing a Simple Polytope from its Graph

    Full text link
    Blind and Mani (1987) proved that the entire combinatorial structure (the vertex-facet incidences) of a simple convex polytope is determined by its abstract graph. Their proof is not constructive. Kalai (1988) found a short, elegant, and algorithmic proof of that result. However, his algorithm has always exponential running time. We show that the problem to reconstruct the vertex-facet incidences of a simple polytope P from its graph can be formulated as a combinatorial optimization problem that is strongly dual to the problem of finding an abstract objective function on P (i.e., a shelling order of the facets of the dual polytope of P). Thereby, we derive polynomial certificates for both the vertex-facet incidences as well as for the abstract objective functions in terms of the graph of P. The paper is a variation on joint work with Michael Joswig and Friederike Koerner (2001).Comment: 14 page

    Computing the bounded subcomplex of an unbounded polyhedron

    Full text link
    We study efficient combinatorial algorithms to produce the Hasse diagram of the poset of bounded faces of an unbounded polyhedron, given vertex-facet incidences. We also discuss the special case of simple polyhedra and present computational results.Comment: 16 page

    GRAVITY: the Calibration Unit

    Full text link
    We present in this paper the design and characterisation of a new sub-system of the VLTI 2nd generation instrument GRAVITY: the Calibration Unit. The Calibration Unit provides all functions to test and calibrate the beam combiner instrument: it creates two artificial stars on four beams, and dispose of four delay lines with an internal metrology. It also includes artificial stars for the tip-tilt and pupil guiding systems, as well as four metrology pick-up diodes, for tests and calibration of the corresponding sub-systems. The calibration unit also hosts the reference targets to align GRAVITY to the VLTI, and the safety shutters to avoid the metrology light to propagate in the VLTI-lab. We present the results of the characterisation and validtion of these differrent sub-units.Comment: 12 pages, 11 figures. Proceeding of SPIE 9146 "Optical and Infrared Interferometry IV

    The GRAVITY metrology system: modeling a metrology in optical fibers

    Full text link
    GRAVITY is the second generation VLT Interferometer (VLTI) instrument for high-precision narrow-angle astrometry and phase-referenced interferometric imaging. The laser metrology system of GRAVITY is at the heart of its astrometric mode, which must measure the distance of 2 stars with a precision of 10 micro-arcseconds. This means the metrology has to measure the optical path difference between the two beam combiners of GRAVITY to a level of 5 nm. The metrology design presents some non-common paths that have consequently to be stable at a level of 1 nm. Otherwise they would impact the performance of GRAVITY. The various tests we made in the past on the prototype give us hints on the components responsible for this error, and on their respective contribution to the total error. It is however difficult to assess their exact origin from only OPD measurements, and therefore, to propose a solution to this problem. In this paper, we present the results of a semi-empirical modeling of the fibered metrology system, relying on theoretical basis, as well as on characterisations of key components. The modeling of the metrology system regarding various effects, e.g., temperature, waveguide heating or mechanical stress, will help us to understand how the metrology behave. The goals of this modeling are to 1) model the test set-ups and reproduce the measurements (as a validation of the modeling), 2) determine the origin of the non-common path errors, and 3) propose modifications to the current metrology design to reach the required 1nm stability.Comment: 20 pages, 19 figures. Proceeding of SPIE 9146 "Optical and Infrared Interferometry IV

    Estimating the phase in ground-based interferometry: performance comparison between single-mode and multimode schemes

    Full text link
    In this paper we compare the performance of multi and single-mode interferometry for the estimation of the phase of the complex visibility. We provide a theoretical description of the interferometric signal which enables to derive the phase error in presence of detector, photon and atmospheric noises, for both multi and single-mode cases. We show that, despite the loss of flux occurring when injecting the light in the single-mode component (i.e. single-mode fibers, integrated optics), the spatial filtering properties of such single-mode devices often enable higher performance than multimode concepts. In the high flux regime speckle noise dominated, single-mode interferometry is always more efficient, and its performance is significantly better when the correction provided by adaptive optics becomes poor, by a factor of 2 and more when the Strehl ratio is lower than 10%. In low light level cases (detector noise regime), multimode interferometry reaches better performance, yet the gain never exceeds 20%, which corresponds to the percentage of photon loss due to the injection in the guides. Besides, we demonstrate that single-mode interferometry is also more robust to the turbulence in both cases of fringe tracking and phase referencing, at the exception of narrow field of views (<1 arcsec).Comment: 9 pages (+ 11 online material appendices) -- 8 Figures. Accepted in A&

    Polytopality and Cartesian products of graphs

    Full text link
    We study the question of polytopality of graphs: when is a given graph the graph of a polytope? We first review the known necessary conditions for a graph to be polytopal, and we provide several families of graphs which satisfy all these conditions, but which nonetheless are not graphs of polytopes. Our main contribution concerns the polytopality of Cartesian products of non-polytopal graphs. On the one hand, we show that products of simple polytopes are the only simple polytopes whose graph is a product. On the other hand, we provide a general method to construct (non-simple) polytopal products whose factors are not polytopal.Comment: 21 pages, 10 figure
    corecore