462 research outputs found

    Closing the Gap Between Short and Long XORs for Model Counting

    Full text link
    Many recent algorithms for approximate model counting are based on a reduction to combinatorial searches over random subsets of the space defined by parity or XOR constraints. Long parity constraints (involving many variables) provide strong theoretical guarantees but are computationally difficult. Short parity constraints are easier to solve but have weaker statistical properties. It is currently not known how long these parity constraints need to be. We close the gap by providing matching necessary and sufficient conditions on the required asymptotic length of the parity constraints. Further, we provide a new family of lower bounds and the first non-trivial upper bounds on the model count that are valid for arbitrarily short XORs. We empirically demonstrate the effectiveness of these bounds on model counting benchmarks and in a Satisfiability Modulo Theory (SMT) application motivated by the analysis of contingency tables in statistics.Comment: The 30th Association for the Advancement of Artificial Intelligence (AAAI-16) Conferenc

    Blunt Force Trauma to the Ribs: Creating Predictive Models

    Get PDF
    Forensic anthropologists receive more requests for trauma analysis than any other aspect of the biological profile. Blunt force trauma to the ribs is some of the most common trauma recorded in a medical examiner’s setting, however the structural complexity of ribs make it difficult to move beyond descriptive documentation of injuries. The purpose of this study is to identify common rib fracture patterns, influential variables, and provide probabilistic statements to guide rib fracture interpretations. A sample of 1,415 deceased individuals with known blunt force trauma to the torso were collected from four geographically diverse medical examiner offices. Demographic data and fracture variables were recorded per individual. Frequency distributions, chi-squared tests, Kruskal-Wallis tests of independence, Dunn’s tests, and multiple correspondence analysis were employed to understand variable relationships. Conditional probabilities were calculated to provide probabilistic statements. Additionally, random forest analysis was conducted to classify location and type of fracture based on covariates. A total of 24, 853 fractures were recorded. The most common fractures were displaced and simple fractures on ribs three through seven in the anterolateral and posterolateral locations. The less common fracture patterns revealed significant relationships with demographic data and provided empirical evidence for previously untested statements. BMI had a significant relationship with location, such that fractures were more frequently recorded in lower ribs in individuals with a BMI category of obese. Age had a significant relationship with fracture type and fracture location in all analyses; younger individuals were more likely to have incomplete fractures and incur fractures anteriorly, and older individuals were more likely to have multi-fragmentary fractures. The current study indicates that rib fracture types and location are dependent on the demographics of the individual. Demographics, such as age and health of the individual inform the material properties and structural geometry of bone, which is how bone biomechanics are recommended to be incorporated into trauma analysis. Furthermore, the results from this research can be applied to motor vehicle safety research, experimental research avenues, and bioarcheological trauma analysis. Future rib fracture research should focus on including a more holistic view of an individual during the interpretation of fracture patterns

    Functional Sensory Representations of Natural Stimuli: the Case of Spatial Hearing

    Get PDF
    In this thesis I attempt to explain mechanisms of neuronal coding in the auditory system as a form of adaptation to statistics of natural stereo sounds. To this end I analyse recordings of real-world auditory environments and construct novel statistical models of these data. I further compare regularities present in natural stimuli with known, experimentally observed neuronal mechanisms of spatial hearing. In a more general perspective, I use binaural auditory system as a starting point to consider the notion of function implemented by sensory neurons. In particular I argue for two, closely-related tenets: 1. The function of sensory neurons can not be fully elucidated without understanding statistics of natural stimuli they process. 2. Function of sensory representations is determined by redundancies present in the natural sensory environment. I present the evidence in support of the first tenet by describing and analysing marginal statistics of natural binaural sound. I compare observed, empirical distributions with knowledge from reductionist experiments. Such comparison allows to argue that the complexity of the spatial hearing task in the natural environment is much higher than analytic, physics-based predictions. I discuss the possibility that early brain stem circuits such as LSO and MSO do not \"compute sound localization\" as is often being claimed in the experimental literature. I propose that instead they perform a signal transformation, which constitutes the first step of a complex inference process. To support the second tenet I develop a hierarchical statistical model, which learns a joint sparse representation of amplitude and phase information from natural stereo sounds. I demonstrate that learned higher order features reproduce properties of auditory cortical neurons, when probed with spatial sounds. Reproduced aspects were hypothesized to be a manifestation of a fine-tuned computation specific to the sound-localization task. Here it is demonstrated that they rather reflect redundancies present in the natural stimulus. Taken together, results presented in this thesis suggest that efficient coding is a strategy useful for discovering structures (redundancies) in the input data. Their meaning has to be determined by the organism via environmental feedback

    Recessions Or Partisanship: What Explains Climate Skepticism in the U.S.?

    Get PDF
    This paper investigates the variations in public mood pertaining to climate skepticism and attempts to empirically assess whether economic recessions or partisanship help explain aggregate-level trends and movements across a 16-year time horizon. Public survey data from the iPoll and Gallup Organization were used to construct the Climate Change Skeptic Index (CCSI) that served as a proxy to capture public opinion trends in skepticism across the U.S. A two-part vector autoregressive model suggests that while economic recessions might be causally linked to climate skepticism, partisanship plays a more influential role in explaining it over time. The key result is that holding all included variables constant, anti-climate change statements by Republican Congresspersons made three quarters ago raise the CCSI by 0.17 percentage points on average in the current quarter

    Pyramid: Enhancing Selectivity in Big Data Protection with Count Featurization

    Full text link
    Protecting vast quantities of data poses a daunting challenge for the growing number of organizations that collect, stockpile, and monetize it. The ability to distinguish data that is actually needed from data collected "just in case" would help these organizations to limit the latter's exposure to attack. A natural approach might be to monitor data use and retain only the working-set of in-use data in accessible storage; unused data can be evicted to a highly protected store. However, many of today's big data applications rely on machine learning (ML) workloads that are periodically retrained by accessing, and thus exposing to attack, the entire data store. Training set minimization methods, such as count featurization, are often used to limit the data needed to train ML workloads to improve performance or scalability. We present Pyramid, a limited-exposure data management system that builds upon count featurization to enhance data protection. As such, Pyramid uniquely introduces both the idea and proof-of-concept for leveraging training set minimization methods to instill rigor and selectivity into big data management. We integrated Pyramid into Spark Velox, a framework for ML-based targeting and personalization. We evaluate it on three applications and show that Pyramid approaches state-of-the-art models while training on less than 1% of the raw data
    • …
    corecore