716 research outputs found

    Semantic bottleneck for computer vision tasks

    Full text link
    This paper introduces a novel method for the representation of images that is semantic by nature, addressing the question of computation intelligibility in computer vision tasks. More specifically, our proposition is to introduce what we call a semantic bottleneck in the processing pipeline, which is a crossing point in which the representation of the image is entirely expressed with natural language , while retaining the efficiency of numerical representations. We show that our approach is able to generate semantic representations that give state-of-the-art results on semantic content-based image retrieval and also perform very well on image classification tasks. Intelligibility is evaluated through user centered experiments for failure detection

    Hierarchical ResNeXt Models for Breast Cancer Histology Image Classification

    Full text link
    Microscopic histology image analysis is a cornerstone in early detection of breast cancer. However these images are very large and manual analysis is error prone and very time consuming. Thus automating this process is in high demand. We proposed a hierarchical system of convolutional neural networks (CNN) that classifies automatically patches of these images into four pathologies: normal, benign, in situ carcinoma and invasive carcinoma. We evaluated our system on the BACH challenge dataset of image-wise classification and a small dataset that we used to extend it. Using a train/test split of 75%/25%, we achieved an accuracy rate of 0.99 on the test split for the BACH dataset and 0.96 on that of the extension. On the test of the BACH challenge, we've reached an accuracy of 0.81 which rank us to the 8th out of 51 teams

    Apollo experience report: Development of guidance targeting techniques for the command module and launch vehicle

    Get PDF
    The development of the guidance targeting techniques for the Apollo command module and launch vehicle is discussed for four types of maneuvers: (1) translunar injection, (2) translunar midcourse, (3) lunar orbit insertion, and (4) return to earth. The development of real-time targeting programs for these maneuvers and the targeting procedures represented are discussed. The material is intended to convey historically the development of the targeting techniques required to meet the defined target objectives and to illustrate the solutions to problems encountered during that development

    Are You Tampering With My Data?

    Full text link
    We propose a novel approach towards adversarial attacks on neural networks (NN), focusing on tampering the data used for training instead of generating attacks on trained models. Our network-agnostic method creates a backdoor during training which can be exploited at test time to force a neural network to exhibit abnormal behaviour. We demonstrate on two widely used datasets (CIFAR-10 and SVHN) that a universal modification of just one pixel per image for all the images of a class in the training set is enough to corrupt the training procedure of several state-of-the-art deep neural networks causing the networks to misclassify any images to which the modification is applied. Our aim is to bring to the attention of the machine learning community, the possibility that even learning-based methods that are personally trained on public datasets can be subject to attacks by a skillful adversary.Comment: 18 page

    'Part'ly first among equals: Semantic part-based benchmarking for state-of-the-art object recognition systems

    Full text link
    An examination of object recognition challenge leaderboards (ILSVRC, PASCAL-VOC) reveals that the top-performing classifiers typically exhibit small differences amongst themselves in terms of error rate/mAP. To better differentiate the top performers, additional criteria are required. Moreover, the (test) images, on which the performance scores are based, predominantly contain fully visible objects. Therefore, `harder' test images, mimicking the challenging conditions (e.g. occlusion) in which humans routinely recognize objects, need to be utilized for benchmarking. To address the concerns mentioned above, we make two contributions. First, we systematically vary the level of local object-part content, global detail and spatial context in images from PASCAL VOC 2010 to create a new benchmarking dataset dubbed PPSS-12. Second, we propose an object-part based benchmarking procedure which quantifies classifiers' robustness to a range of visibility and contextual settings. The benchmarking procedure relies on a semantic similarity measure that naturally addresses potential semantic granularity differences between the category labels in training and test datasets, thus eliminating manual mapping. We use our procedure on the PPSS-12 dataset to benchmark top-performing classifiers trained on the ILSVRC-2012 dataset. Our results show that the proposed benchmarking procedure enables additional differentiation among state-of-the-art object classifiers in terms of their ability to handle missing content and insufficient object detail. Given this capability for additional differentiation, our approach can potentially supplement existing benchmarking procedures used in object recognition challenge leaderboards.Comment: Extended version of our ACCV-2016 paper. Author formatting modifie

    Role of dynamic Jahn-Teller distortions in Na2C60 and Na2CsC60 studied by NMR

    Full text link
    Through 13C NMR spin lattice relaxation (T1) measurements in cubic Na2C60, we detect a gap in its electronic excitations, similar to that observed in tetragonal A4C60. This establishes that Jahn-Teller distortions (JTD) and strong electronic correlations must be considered to understand the behaviour of even electron systems, regardless of the structure. Furthermore, in metallic Na2CsC60, a similar contribution to T1 is also detected for 13C and 133Cs NMR, implying the occurence of excitations typical of JT distorted C60^{2-} (or equivalently C60^{4-}). This supports the idea that dynamic JTD can induce attractive electronic interactions in odd electron systems.Comment: 3 figure

    Transport control by coherent zonal flows in the core/edge transitional regime

    Get PDF
    3D Braginskii turbulence simulations show that the energy flux in the core/edge transition region of a tokamak is strongly modulated - locally and on average - by radially propagating, nearly coherent sinusoidal or solitary zonal flows. The flows are geodesic acoustic modes (GAM), which are primarily driven by the Stringer-Winsor term. The flow amplitude together with the average anomalous transport sensitively depend on the GAM frequency and on the magnetic curvature acting on the flows, which could be influenced in a real tokamak, e.g., by shaping the plasma cross section. The local modulation of the turbulence by the flows and the excitation of the flows are due to wave-kinetic effects, which have been studied for the first time in a turbulence simulation.Comment: 5 pages, 5 figures, submitted to PR

    Detecting and tracing building occupants to optimize process control

    Get PDF

    Comparison of high versus low frequency cerebral physiology for cerebrovascular reactivity assessment in traumatic brain injury: a multi-center pilot study

    Get PDF
    Current accepted cerebrovascular reactivity indices suffer from the need of high frequency data capture and export for post-acquisition processing. The role for minute-by-minute data in cerebrovascular reactivity monitoring remains uncertain. The goal was to explore the statistical time-series relationships between intra-cranial pressure (ICP), mean arterial pressure (MAP) and pressure reactivity index (PRx) using both 10-s and minute data update frequency in TBI. Prospective data from 31 patients from 3 centers with moderate/severe TBI and high-frequency archived physiology were reviewed. Both 10-s by 10-s and minute-by-minute mean values were derived for ICP and MAP for each patient. Similarly, PRx was derived using 30 consecutive 10-s data points, updated every minute. While long-PRx (L-PRx) was derived via similar methodology using minute-by-minute data, with L-PRx derived using various window lengths (5, 10, 20, 30, 40, and 60 min; denoted L-PRx_5, etc.). Time-series autoregressive integrative moving average (ARIMA) and vector autoregressive integrative moving average (VARIMA) models were created to analyze the relationship of these parameters over time. ARIMA modelling, Granger causality testing and VARIMA impulse response function (IRF) plotting demonstrated that similar information is carried in minute mean ICP and MAP data, compared to 10-s mean slow-wave ICP and MAP data. Shorter window L-PRx variants, such as L-PRx_5, appear to have a similar ARIMA structure, have a linear association with PRx and display moderate-to-strong correlations (r ~ 0.700, p Peer reviewe

    PlaNet - Photo Geolocation with Convolutional Neural Networks

    Full text link
    Is it possible to build a system to determine the location where a photo was taken using just its pixels? In general, the problem seems exceptionally difficult: it is trivial to construct situations where no location can be inferred. Yet images often contain informative cues such as landmarks, weather patterns, vegetation, road markings, and architectural details, which in combination may allow one to determine an approximate location and occasionally an exact location. Websites such as GeoGuessr and View from your Window suggest that humans are relatively good at integrating these cues to geolocate images, especially en-masse. In computer vision, the photo geolocation problem is usually approached using image retrieval methods. In contrast, we pose the problem as one of classification by subdividing the surface of the earth into thousands of multi-scale geographic cells, and train a deep network using millions of geotagged images. While previous approaches only recognize landmarks or perform approximate matching using global image descriptors, our model is able to use and integrate multiple visible cues. We show that the resulting model, called PlaNet, outperforms previous approaches and even attains superhuman levels of accuracy in some cases. Moreover, we extend our model to photo albums by combining it with a long short-term memory (LSTM) architecture. By learning to exploit temporal coherence to geolocate uncertain photos, we demonstrate that this model achieves a 50% performance improvement over the single-image model
    • …
    corecore