930 research outputs found

    ESTSS at 20 years: "a phoenix gently rising from a lava flow of European trauma"

    Get PDF
    Roderick J. Ørner, who was President between 1997 and 1999, traces the phoenix-like origins of the European Society for Traumatic Stress Studies (ESTSS) from an informal business meeting called during the 1st European Conference on Traumatic Stress (ECOTS) in 1987 to its emergence into a formally constituted society. He dwells on the challenges of tendering a trauma society within a continent where trauma has been and remains endemic. ESTSS successes are noted along with a number of personal reflections on activities that give rise to concern for the present as well as its future prospects. Denial of survivors' experiences and turning away from survivors' narratives by reframing their experiences to accommodate helpers' theory-driven imperatives are viewed with alarm. Arguments are presented for making human rights, memory, and ethics core elements of a distinctive European psycho traumatology, which will secure current ESTSS viability and future integrity

    Spectral Sparsification and Regret Minimization Beyond Matrix Multiplicative Updates

    Full text link
    In this paper, we provide a novel construction of the linear-sized spectral sparsifiers of Batson, Spielman and Srivastava [BSS14]. While previous constructions required Ω(n4)\Omega(n^4) running time [BSS14, Zou12], our sparsification routine can be implemented in almost-quadratic running time O(n2+ε)O(n^{2+\varepsilon}). The fundamental conceptual novelty of our work is the leveraging of a strong connection between sparsification and a regret minimization problem over density matrices. This connection was known to provide an interpretation of the randomized sparsifiers of Spielman and Srivastava [SS11] via the application of matrix multiplicative weight updates (MWU) [CHS11, Vis14]. In this paper, we explain how matrix MWU naturally arises as an instance of the Follow-the-Regularized-Leader framework and generalize this approach to yield a larger class of updates. This new class allows us to accelerate the construction of linear-sized spectral sparsifiers, and give novel insights on the motivation behind Batson, Spielman and Srivastava [BSS14]

    Scalable and Interpretable One-class SVMs with Deep Learning and Random Fourier features

    Full text link
    One-class support vector machine (OC-SVM) for a long time has been one of the most effective anomaly detection methods and extensively adopted in both research as well as industrial applications. The biggest issue for OC-SVM is yet the capability to operate with large and high-dimensional datasets due to optimization complexity. Those problems might be mitigated via dimensionality reduction techniques such as manifold learning or autoencoder. However, previous work often treats representation learning and anomaly prediction separately. In this paper, we propose autoencoder based one-class support vector machine (AE-1SVM) that brings OC-SVM, with the aid of random Fourier features to approximate the radial basis kernel, into deep learning context by combining it with a representation learning architecture and jointly exploit stochastic gradient descent to obtain end-to-end training. Interestingly, this also opens up the possible use of gradient-based attribution methods to explain the decision making for anomaly detection, which has ever been challenging as a result of the implicit mappings between the input space and the kernel space. To the best of our knowledge, this is the first work to study the interpretability of deep learning in anomaly detection. We evaluate our method on a wide range of unsupervised anomaly detection tasks in which our end-to-end training architecture achieves a performance significantly better than the previous work using separate training.Comment: Accepted at European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD) 201

    Private Incremental Regression

    Full text link
    Data is continuously generated by modern data sources, and a recent challenge in machine learning has been to develop techniques that perform well in an incremental (streaming) setting. In this paper, we investigate the problem of private machine learning, where as common in practice, the data is not given at once, but rather arrives incrementally over time. We introduce the problems of private incremental ERM and private incremental regression where the general goal is to always maintain a good empirical risk minimizer for the history observed under differential privacy. Our first contribution is a generic transformation of private batch ERM mechanisms into private incremental ERM mechanisms, based on a simple idea of invoking the private batch ERM procedure at some regular time intervals. We take this construction as a baseline for comparison. We then provide two mechanisms for the private incremental regression problem. Our first mechanism is based on privately constructing a noisy incremental gradient function, which is then used in a modified projected gradient procedure at every timestep. This mechanism has an excess empirical risk of d\approx\sqrt{d}, where dd is the dimensionality of the data. While from the results of [Bassily et al. 2014] this bound is tight in the worst-case, we show that certain geometric properties of the input and constraint set can be used to derive significantly better results for certain interesting regression problems.Comment: To appear in PODS 201

    Adaptive Stochastic Primal-Dual Coordinate Descent for Separable Saddle Point Problems

    Get PDF
    We consider a generic convex-concave saddle point problem with separable structure, a form that covers a wide-ranged machine learning applications. Under this problem structure, we follow the framework of primal-dual updates for saddle point problems, and incorporate stochastic block coordinate descent with adaptive stepsize into this framework. We theoretically show that our proposal of adaptive stepsize potentially achieves a sharper linear convergence rate compared with the existing methods. Additionally, since we can select "mini-batch" of block coordinates to update, our method is also amenable to parallel processing for large-scale data. We apply the proposed method to regularized empirical risk minimization and show that it performs comparably or, more often, better than state-of-the-art methods on both synthetic and real-world data sets.Comment: Accepted by ECML/PKDD201

    Robustness and Generalization

    Full text link
    We derive generalization bounds for learning algorithms based on their robustness: the property that if a testing sample is "similar" to a training sample, then the testing error is close to the training error. This provides a novel approach, different from the complexity or stability arguments, to study generalization of learning algorithms. We further show that a weak notion of robustness is both sufficient and necessary for generalizability, which implies that robustness is a fundamental property for learning algorithms to work

    Stress-Induced Reinstatement of Drug Seeking: 20 Years of Progress

    Get PDF
    In human addicts, drug relapse and craving are often provoked by stress. Since 1995, this clinical scenario has been studied using a rat model of stress-induced reinstatement of drug seeking. Here, we first discuss the generality of stress-induced reinstatement to different drugs of abuse, different stressors, and different behavioral procedures. We also discuss neuropharmacological mechanisms, and brain areas and circuits controlling stress-induced reinstatement of drug seeking. We conclude by discussing results from translational human laboratory studies and clinical trials that were inspired by results from rat studies on stress-induced reinstatement. Our main conclusions are (1) The phenomenon of stress-induced reinstatement, first shown with an intermittent footshock stressor in rats trained to self-administer heroin, generalizes to other abused drugs, including cocaine, methamphetamine, nicotine, and alcohol, and is also observed in the conditioned place preference model in rats and mice. This phenomenon, however, is stressor specific and not all stressors induce reinstatement of drug seeking. (2) Neuropharmacological studies indicate the involvement of corticotropin-releasing factor (CRF), noradrenaline, dopamine, glutamate, kappa/dynorphin, and several other peptide and neurotransmitter systems in stress-induced reinstatement. Neuropharmacology and circuitry studies indicate the involvement of CRF and noradrenaline transmission in bed nucleus of stria terminalis and central amygdala, and dopamine, CRF, kappa/dynorphin, and glutamate transmission in other components of the mesocorticolimbic dopamine system (ventral tegmental area, medial prefrontal cortex, orbitofrontal cortex, and nucleus accumbens). (3) Translational human laboratory studies and a recent clinical trial study show the efficacy of alpha-2 adrenoceptor agonists in decreasing stress-induced drug craving and stress-induced initial heroin lapse
    corecore