529 research outputs found

    Shuttle electrical environment

    Get PDF
    Part of an AFGL payload flown on the STS-4 mission consisted of experiments to measure in-situ electric fields, electron densities, and vehicle charging. During this flight some 11 hours of data were acquired ranging from 5 minute snapshots up to continuous half-orbits. These experiments are described and results presented for such vehicle induced events as a main engine burn, thruster firings and water dumps in addition to undisturbed periods. The main characteristic of all the vehicle induced events is shown to be an enhancement in the low frequency noise (less than 2 kHz), in both the electrostatic and electron irregularity (delta N/N) spectra. The non-event results indicate that the electrostatic broadband emissions show a white noise characteristic in the low frequency range up to 2 kHz at an amplitude of 10 db above the shuttle design specification limit, falling below that limit above 10 kHz. The vehicle potential remained within the range of -3 to +1 volt throughout the flight which exhibits normal behavior for a satellite in a low equatorial orbit

    Phone a Friend or Ask Alexa? Children’s Trust in Voice-Activated Devices

    Get PDF
    Voice-activated devices such as Google Home, Siri, and Alexa are in many homes and children are interacting with these devices. It is unclear if they treat these devices the way they treat human informants. Children prefer human informants that are reliable and familiar. This study examined whether children believe voice-activated devices provide accurate information. Participants included 40 4- and 5-year-olds and 40 7- and 8-year-olds. Children were introduced to two informants: the experimenter’s good friend and the experimenter’s new device. Children heard questions about personal information (e.g., the experimenter’s favorite color), facts that do not change (e.g., the color of a kiwano fruit), and timely information (e.g., which state had the most rain yesterday). After the informant provided an answer, the child indicated whether the answer was correct. Older children were significantly more likely to trust the device’s stable fact responses and the human informant’s personal fact responses. Surprisingly, younger children did not show greater trust for either informant for stable facts, but were significantly more likely to trust personal facts given by the device. These findings suggest that younger children have greater difficulty than older children trusting the appropriate informant, and thus need more guidance from adults to understand and use voice-activated devices.https://ir.library.louisville.edu/uars/1036/thumbnail.jp

    Kepler Presearch Data Conditioning II - A Bayesian Approach to Systematic Error Correction

    Full text link
    With the unprecedented photometric precision of the Kepler Spacecraft, significant systematic and stochastic errors on transit signal levels are observable in the Kepler photometric data. These errors, which include discontinuities, outliers, systematic trends and other instrumental signatures, obscure astrophysical signals. The Presearch Data Conditioning (PDC) module of the Kepler data analysis pipeline tries to remove these errors while preserving planet transits and other astrophysically interesting signals. The completely new noise and stellar variability regime observed in Kepler data poses a significant problem to standard cotrending methods such as SYSREM and TFA. Variable stars are often of particular astrophysical interest so the preservation of their signals is of significant importance to the astrophysical community. We present a Bayesian Maximum A Posteriori (MAP) approach where a subset of highly correlated and quiet stars is used to generate a cotrending basis vector set which is in turn used to establish a range of "reasonable" robust fit parameters. These robust fit parameters are then used to generate a Bayesian Prior and a Bayesian Posterior Probability Distribution Function (PDF) which when maximized finds the best fit that simultaneously removes systematic effects while reducing the signal distortion and noise injection which commonly afflicts simple least-squares (LS) fitting. A numerical and empirical approach is taken where the Bayesian Prior PDFs are generated from fits to the light curve distributions themselves.Comment: 43 pages, 21 figures, Submitted for publication in PASP. Also see companion paper "Kepler Presearch Data Conditioning I - Architecture and Algorithms for Error Correction in Kepler Light Curves" by Martin C. Stumpe, et a

    Kepler Presearch Data Conditioning I - Architecture and Algorithms for Error Correction in Kepler Light Curves

    Full text link
    Kepler provides light curves of 156,000 stars with unprecedented precision. However, the raw data as they come from the spacecraft contain significant systematic and stochastic errors. These errors, which include discontinuities, systematic trends, and outliers, obscure the astrophysical signals in the light curves. To correct these errors is the task of the Presearch Data Conditioning (PDC) module of the Kepler data analysis pipeline. The original version of PDC in Kepler did not meet the extremely high performance requirements for the detection of miniscule planet transits or highly accurate analysis of stellar activity and rotation. One particular deficiency was that astrophysical features were often removed as a side-effect to removal of errors. In this paper we introduce the completely new and significantly improved version of PDC which was implemented in Kepler SOC 8.0. This new PDC version, which utilizes a Bayesian approach for removal of systematics, reliably corrects errors in the light curves while at the same time preserving planet transits and other astrophysically interesting signals. We describe the architecture and the algorithms of this new PDC module, show typical errors encountered in Kepler data, and illustrate the corrections using real light curve examples.Comment: Submitted to PASP. Also see companion paper "Kepler Presearch Data Conditioning II - A Bayesian Approach to Systematic Error Correction" by Jeff C. Smith et a

    Photometric Analysis in the Kepler Science Operations Center Pipeline

    Get PDF
    We describe the Photometric Analysis (PA) software component and its context in the Kepler Science Operations Center (SOC) pipeline. The primary tasks of this module are to compute the photometric flux and photocenters (centroids) for over 160,000 long cadence (~thirty minute) and 512 short cadence (~one minute) stellar targets from the calibrated pixels in their respective apertures. We discuss the science algorithms for long and short cadence PA: cosmic ray cleaning; background estimation and removal; aperture photometry; and flux-weighted centroiding. We discuss the end-to-end propagation of uncertainties for the science algorithms. Finally, we present examples of photometric apertures, raw flux light curves, and centroid time series from Kepler flight data. PA light curves, centroid time series, and barycentric timestamp corrections are exported to the Multi-mission Archive at Space Telescope [Science Institute] (MAST) and are made available to the general public in accordance with the NASA/Kepler data release policy

    Lessons learned from the introduction of autonomous monitoring to the EUVE science operations center

    Get PDF
    The University of California at Berkeley's (UCB) Center for Extreme Ultraviolet Astrophysics (CEA), in conjunction with NASA's Ames Research Center (ARC), has implemented an autonomous monitoring system in the Extreme Ultraviolet Explorer (EUVE) science operations center (ESOC). The implementation was driven by a need to reduce operations costs and has allowed the ESOC to move from continuous, three-shift, human-tended monitoring of the science payload to a one-shift operation in which the off shifts are monitored by an autonomous anomaly detection system. This system includes Eworks, an artificial intelligence (AI) payload telemetry monitoring package based on RTworks, and Epage, an automatic paging system to notify ESOC personnel of detected anomalies. In this age of shrinking NASA budgets, the lessons learned on the EUVE project are useful to other NASA missions looking for ways to reduce their operations budgets. The process of knowledge capture, from the payload controllers for implementation in an expert system, is directly applicable to any mission considering a transition to autonomous monitoring in their control center. The collaboration with ARC demonstrates how a project with limited programming resources can expand the breadth of its goals without incurring the high cost of hiring additional, dedicated programmers. This dispersal of expertise across NASA centers allows future missions to easily access experts for collaborative efforts of their own. Even the criterion used to choose an expert system has widespread impacts on the implementation, including the completion time and the final cost. In this paper we discuss, from inception to completion, the areas where our experiences in moving from three shifts to one shift may offer insights for other NASA missions

    Detection of Potential Transit Signals in the First Three Quarters of Kepler Mission Data

    Full text link
    We present the results of a search for potential transit signals in the first three quarters of photometry data acquired by the Kepler Mission. The targets of the search include 151,722 stars which were observed over the full interval and an additional 19,132 stars which were observed for only 1 or 2 quarters. From this set of targets we find a total of 5,392 detections which meet the Kepler detection criteria: those criteria are periodicity of the signal, an acceptable signal-to-noise ratio, and a composition test which rejects spurious detections which contain non-physical combinations of events. The detected signals are dominated by events with relatively low signal-to-noise ratio and by events with relatively short periods. The distribution of estimated transit depths appears to peak in the range between 40 and 100 parts per million, with a few detections down to fewer than 10 parts per million. The detected signals are compared to a set of known transit events in the Kepler field of view which were derived by a different method using a longer data interval; the comparison shows that the current search correctly identified 88.1% of the known events. A tabulation of the detected transit signals, examples which illustrate the analysis and detection process, a discussion of future plans and open, potentially fruitful, areas of further research are included
    • …
    corecore