2,890 research outputs found

    Graph Summarization

    Full text link
    The continuous and rapid growth of highly interconnected datasets, which are both voluminous and complex, calls for the development of adequate processing and analytical techniques. One method for condensing and simplifying such datasets is graph summarization. It denotes a series of application-specific algorithms designed to transform graphs into more compact representations while preserving structural patterns, query answers, or specific property distributions. As this problem is common to several areas studying graph topologies, different approaches, such as clustering, compression, sampling, or influence detection, have been proposed, primarily based on statistical and optimization methods. The focus of our chapter is to pinpoint the main graph summarization methods, but especially to focus on the most recent approaches and novel research trends on this topic, not yet covered by previous surveys.Comment: To appear in the Encyclopedia of Big Data Technologie

    Simultaneous Measurement Imputation and Outcome Prediction for Achilles Tendon Rupture Rehabilitation

    Full text link
    Achilles Tendon Rupture (ATR) is one of the typical soft tissue injuries. Rehabilitation after such a musculoskeletal injury remains a prolonged process with a very variable outcome. Accurately predicting rehabilitation outcome is crucial for treatment decision support. However, it is challenging to train an automatic method for predicting the ATR rehabilitation outcome from treatment data, due to a massive amount of missing entries in the data recorded from ATR patients, as well as complex nonlinear relations between measurements and outcomes. In this work, we design an end-to-end probabilistic framework to impute missing data entries and predict rehabilitation outcomes simultaneously. We evaluate our model on a real-life ATR clinical cohort, comparing with various baselines. The proposed method demonstrates its clear superiority over traditional methods which typically perform imputation and prediction in two separate stages

    Estimation of the vertical wavelength of atmospheric gravity waves from airglow imagery

    Get PDF
    Abstract In the summer of 2010, two imagers were installed in New Mexico with the objective of making stereoscopic observations of atmospheric gravity waves (AGWs). As AGWs propagate vertically, they spatially perturb the airglow emission layers in all three dimensions. Estimates of the vertical wavelength, horizontal wavelength, and the intrinsic frequency are needed to characterize an AGW and quantify its effects on upper atmospheric dynamics. The dispersion relation describes the relationship between vertical and horizontal wavelengths as a function of the intrinsic frequency. Thus, any two of the three aforementioned parameters can be used to determine the third. Mesospheric winds are hard to measure and consequently the intrinsic frequency is difficult to estimate. However, the horizontal wavelength can be directly measured from airglow imagery once the three-dimensional imager field of view is projected onto the two-dimensional image plane. This thesis presents a method to estimate the vertical wavelength using an airglow perturbation model proposed by Anderson et al. (2009). The model is subsequently validated using the observations from ground-based imagers installed in New Mexico. Abstract The perturbed airglow is modeled as a quasi-monochromatic wave and thus, it can be characterized using only a few parameters, one of which is the vertical wavelength. Because the vertical wavelength is embedded in both the phase and the magnitude of this model, two values of the vertical wavelength are estimated by applying two different parameter estimation techniques on the phase and magnitude. The estimation of the vertical wavelength from the phase of the model entails solving an overdetermined system of linear equations by minimizing the sum of the squared residuals. This estimate is then compared to that obtained by iteratively finding the best approximation to the roots of a function, representing the magnitude of the perturbation model. These two techniques are applied on three nights in 2010, and the estimates for the vertical wavelength match to within a few kilometers. Thus, the perturbation model is validated using real data

    Neural ODEs with stochastic vector field mixtures

    Full text link
    It was recently shown that neural ordinary differential equation models cannot solve fundamental and seemingly straightforward tasks even with high-capacity vector field representations. This paper introduces two other fundamental tasks to the set that baseline methods cannot solve, and proposes mixtures of stochastic vector fields as a model class that is capable of solving these essential problems. Dynamic vector field selection is of critical importance for our model, and our approach is to propagate component uncertainty over the integration interval with a technique based on forward filtering. We also formalise several loss functions that encourage desirable properties on the trajectory paths, and of particular interest are those that directly encourage fewer expected function evaluations. Experimentally, we demonstrate that our model class is capable of capturing the natural dynamics of human behaviour; a notoriously volatile application area. Baseline approaches cannot adequately model this problem

    Computational Methods for Probabilistic Inference of Sector Congestion in Air Traffic Management

    Get PDF
    This article addresses the issue of computing the expected cost functions from a probabilistic model of the air traffic flow and capacity management. The Clenshaw-Curtis quadrature is compared to Monte-Carlo algorithms defined specifically for this problem. By tailoring the algorithms to this model, we reduce the computational burden in order to simulate real instances. The study shows that the Monte-Carlo algorithm is more sensible to the amount of uncertainty in the system, but has the advantage to return a result with the associated accuracy on demand. The performances for both approaches are comparable for the computation of the expected cost of delay and the expected cost of congestion. Finally, this study shows some evidences that the simulation of the proposed probabilistic model is tractable for realistic instances.Comment: Interdisciplinary Science for Innovative Air Traffic Management (2013

    Routing in Mobile Ad-Hoc Networks using Social Tie Strengths and Mobility Plans

    Full text link
    We consider the problem of routing in a mobile ad-hoc network (MANET) for which the planned mobilities of the nodes are partially known a priori and the nodes travel in groups. This situation arises commonly in military and emergency response scenarios. Optimal routes are computed using the most reliable path principle in which the negative logarithm of a node pair's adjacency probability is used as a link weight metric. This probability is estimated using the mobility plan as well as dynamic information captured by table exchanges, including a measure of the social tie strength between nodes. The latter information is useful when nodes deviate from their plans or when the plans are inaccurate. We compare the proposed routing algorithm with the commonly-used optimized link state routing (OLSR) protocol in ns-3 simulations. As the OLSR protocol does not exploit the mobility plans, it relies on link state determination which suffers with increasing mobility. Our simulations show considerably better throughput performance with the proposed approach as compared with OLSR at the expense of increased overhead. However, in the high-throughput regime, the proposed approach outperforms OLSR in terms of both throughput and overhead
    corecore