2,004 research outputs found

    The Application of Metadata Standards to Multimedia in Museums

    Get PDF
    This paper first describes the application of a multi-level indexing approach, based on Dublin Core extensions and the Resource Description Framework (RDF), to a typical museum video. The advantages and disadvantages of this approach are discussed in the context of the requirements of the proposed MPEG-7 ("Multimedia Content Description Interface") standard. The work on SMIL (Synchronized Multimedia Integration Language) by the W3C SYMM working group is then described. Suggestions for how this work can be applied to video metadata are made. Finally a hybrid approach is proposed based on the combined use of Dublin Core and the currently undefined MPEG-7 standard within the RDF which will provide a solution to the problem of satisfying widely differing user requirements

    Case study on quantum convolutional neural network scalability

    Full text link
    One of the crucial tasks in computer science is the processing time reduction of various data types, i.e., images, which is important for different fields -- from medicine and logistics to virtual shopping. Compared to classical computers, quantum computers are capable of parallel data processing, which reduces the data processing time. This quality of quantum computers inspired intensive research of the potential of quantum technologies applicability to real-life tasks. Some progress has already revealed on a smaller volumes of the input data. In this research effort, I aimed to increase the amount of input data (I used images from 2 x 2 to 8 x 8), while reducing the processing time, by way of skipping intermediate measurement steps. The hypothesis was that, for increased input data, the omitting of intermediate measurement steps after each quantum convolution layer will improve output metric results and accelerate data processing. To test the hypothesis, I performed experiments to chose the best activation function and its derivative in each network. The hypothesis was partly confirmed in terms of output mean squared error (MSE) -- it dropped from 0.25 in the result of classical convolutional neural network (CNN) training to 0.23 in the result of quantum convolutional neural network (QCNN) training. In terms of the training time, however, which was 1.5 minutes for CNN and 4 hours 37 minutes in the least lengthy training iteration, the hypothesis was rejected.Comment: 11 pages (without references), 13 figure

    ALICE: The Ultraviolet Imaging Spectrograph aboard the New Horizons Pluto-Kuiper Belt Mission

    Full text link
    The New Horizons ALICE instrument is a lightweight (4.4 kg), low-power (4.4 Watt) imaging spectrograph aboard the New Horizons mission to Pluto/Charon and the Kuiper Belt. Its primary job is to determine the relative abundances of various species in Pluto's atmosphere. ALICE will also be used to search for an atmosphere around Pluto's moon, Charon, as well as the Kuiper Belt Objects (KBOs) that New Horizons hopes to fly by after Pluto-Charon, and it will make UV surface reflectivity measurements of all of these bodies as well. The instrument incorporates an off-axis telescope feeding a Rowland-circle spectrograph with a 520-1870 angstroms spectral passband, a spectral point spread function of 3-6 angstroms FWHM, and an instantaneous spatial field-of-view that is 6 degrees long. Different input apertures that feed the telescope allow for both airglow and solar occultation observations during the mission. The focal plane detector is an imaging microchannel plate (MCP) double delay-line detector with dual solar-blind opaque photocathodes (KBr and CsI) and a focal surface that matches the instrument's 15-cm diameter Rowland-circle. In what follows, we describe the instrument in greater detail, including descriptions of its ground calibration and initial in flight performance.Comment: 24 pages, 29 figures, 2 tables; To appear in a special volume of Space Science Reviews on the New Horizons missio

    Event-based simulation of quantum physics experiments

    Full text link
    We review an event-based simulation approach which reproduces the statistical distributions of wave theory not by requiring the knowledge of the solution of the wave equation of the whole system but by generating detection events one-by-one according to an unknown distribution. We illustrate its applicability to various single photon and single neutron interferometry experiments and to two Bell test experiments, a single-photon Einstein-Podolsky-Rosen experiment employing post-selection for photon pair identification and a single-neutron Bell test interferometry experiment with nearly 100%100\% detection efficiency.Comment: Lectures notes of the Advanced School on Quantum Foundations and Open Quantum Systems, Jo\~ao Pessoa, Brazil, July 2012, edited by T. M. Nieuwenhuizen et al, World Scientific, to appea
    corecore