287 research outputs found

    Massively Parallel Video Networks

    Full text link
    We introduce a class of causal video understanding models that aims to improve efficiency of video processing by maximising throughput, minimising latency, and reducing the number of clock cycles. Leveraging operation pipelining and multi-rate clocks, these models perform a minimal amount of computation (e.g. as few as four convolutional layers) for each frame per timestep to produce an output. The models are still very deep, with dozens of such operations being performed but in a pipelined fashion that enables depth-parallel computation. We illustrate the proposed principles by applying them to existing image architectures and analyse their behaviour on two video tasks: action recognition and human keypoint localisation. The results show that a significant degree of parallelism, and implicitly speedup, can be achieved with little loss in performance.Comment: Fixed typos in densenet model definition in appendi

    Self-Supervised Relative Depth Learning for Urban Scene Understanding

    Full text link
    As an agent moves through the world, the apparent motion of scene elements is (usually) inversely proportional to their depth. It is natural for a learning agent to associate image patterns with the magnitude of their displacement over time: as the agent moves, faraway mountains don't move much; nearby trees move a lot. This natural relationship between the appearance of objects and their motion is a rich source of information about the world. In this work, we start by training a deep network, using fully automatic supervision, to predict relative scene depth from single images. The relative depth training images are automatically derived from simple videos of cars moving through a scene, using recent motion segmentation techniques, and no human-provided labels. This proxy task of predicting relative depth from a single image induces features in the network that result in large improvements in a set of downstream tasks including semantic segmentation, joint road segmentation and car detection, and monocular (absolute) depth estimation, over a network trained from scratch. The improvement on the semantic segmentation task is greater than those produced by any other automatically supervised methods. Moreover, for monocular depth estimation, our unsupervised pre-training method even outperforms supervised pre-training with ImageNet. In addition, we demonstrate benefits from learning to predict (unsupervised) relative depth in the specific videos associated with various downstream tasks. We adapt to the specific scenes in those tasks in an unsupervised manner to improve performance. In summary, for semantic segmentation, we present state-of-the-art results among methods that do not use supervised pre-training, and we even exceed the performance of supervised ImageNet pre-trained models for monocular depth estimation, achieving results that are comparable with state-of-the-art methods

    An experimental multiprocessor system for distributed parallel computations.

    Get PDF
    The availability of low-cost microprocessor chips with efficient instruction sets for specific numerical tasks (signal processors) has been exploited for building a versatile multiprocessor system, consisting of a host minicomputer augmented by a number of joint processors. The host provides a multiuser-multitasking environment and manages system resources and task scheduling. User applications can call upon one or more joint processors for parallel execution of adequately partitioned, computationally intensive numeric operations. Each joint processor has sufficient local memory for storing procedures and data and has access to regions in host memory for shared data. Kernel processes in the host and in the joint processors provide the necessary mechanism for initialization and synchronization of the distributed parallel execution of procedures

    Arsenic induces metabolome remodeling in mature human adipocytes.

    Get PDF
    Human lifetime exposure to arsenic through drinking water, food supply or industrial pollution leads to its accumulation in many organs such as liver, kidneys, lungs or pancreas but also adipose tissue. Recently, population-based studies revealed the association between arsenic exposure and the development of metabolic diseases such as obesity and type 2 diabetes. To shed light on the molecular bases of such association, we determined the concentration that inhibited 17% of cell viability and investigated the effects of arsenic acute exposure on adipose-derived human mesenchymal stem cells differentiated in vitro into mature adipocytes and treated with sodium arsenite (NaAsO <sub>2</sub> , 10 nM to 10 µM). Untargeted metabolomics and gene expression analyses revealed a strong dose-dependent inhibition of lipogenesis and lipolysis induction, reducing the cellular ability to store lipids. These dysregulations were emphasized by the inhibition of the cellular response to insulin, as shown by the perturbation of several genes and metabolites involved in the mentioned biological pathways. Our study highlighted the activation of an adaptive oxidative stress response with the strong induction of metallothioneins and increased glutathione levels in response to arsenic accumulation that could exacerbate the decreased insulin sensitivity of the adipocytes. Arsenic exposure strongly affected the expression of arsenic transporters, responsible for arsenic influx and efflux, and induced a pro-inflammatory state in adipocytes by enhancing the expression of the inflammatory interleukin 6 (IL6). Collectively, our data showed that an acute exposure to low levels of arsenic concentrations alters key adipocyte functions, highlighting its contribution to the development of insulin resistance and the pathogenesis of metabolic disorders

    View-tolerant face recognition and Hebbian learning imply mirror-symmetric neural tuning to head orientation

    Get PDF
    The primate brain contains a hierarchy of visual areas, dubbed the ventral stream, which rapidly computes object representations that are both specific for object identity and relatively robust against identity-preserving transformations like depth-rotations. Current computational models of object recognition, including recent deep learning networks, generate these properties through a hierarchy of alternating selectivity-increasing filtering and tolerance-increasing pooling operations, similar to simple-complex cells operations. While simulations of these models recapitulate the ventral stream's progression from early view-specific to late view-tolerant representations, they fail to generate the most salient property of the intermediate representation for faces found in the brain: mirror-symmetric tuning of the neural population to head orientation. Here we prove that a class of hierarchical architectures and a broad set of biologically plausible learning rules can provide approximate invariance at the top level of the network. While most of the learning rules do not yield mirror-symmetry in the mid-level representations, we characterize a specific biologically-plausible Hebb-type learning rule that is guaranteed to generate mirror-symmetric tuning to faces tuning at intermediate levels of the architecture

    A comparison of two computer-based face identification systems with human perceptions of faces

    Get PDF
    The performance of two different computer systems for representing faces was compared with human ratings of similarity and distinctiveness, and human memory performance, on a specific set of face images. The systems compared were a graphmatching system (e.g. Lades et al., 1993) and coding based on Principal Components Analysis (PCA) of image pixels (e.g. Turk & Pentland, 1991). Replicating other work, the PCA-based system produced very much better performance at recognising faces, and higher correlations with human performance with the same images, when the images were initially standardised using a morphing procedure and separate analysis of "shape" and "shape-free" components then combined. Both the graph-matching and (shape + shape-free) PCA systems were equally able to recognise faces shown with changed expressions, both provided reasonable correlations with human ratings and memory data, and there were also correlations between the facial similarities recorded by each of the computer models. However, comparisons with human similarity ratings of faces with and without the hair visible, and prediction of memory performance with and without alteration in face expressions, suggested that the graph-matching system was better at capturing aspects of the appearance of the face, while the PCA-based system seemed better at capturing aspects of the appearance of specific images of faces

    Deep Learning of Representations: Looking Forward

    Full text link
    Deep learning research aims at discovering learning algorithms that discover multiple levels of distributed representations, with higher levels representing more abstract concepts. Although the study of deep learning has already led to impressive theoretical results, learning algorithms and breakthrough experiments, several challenges lie ahead. This paper proposes to examine some of these challenges, centering on the questions of scaling deep learning algorithms to much larger models and datasets, reducing optimization difficulties due to ill-conditioning or local minima, designing more efficient and powerful inference and sampling procedures, and learning to disentangle the factors of variation underlying the observed data. It also proposes a few forward-looking research directions aimed at overcoming these challenges

    The taper of cast post preparation measured using innovative image processing technique

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>No documentation in the literature about taper of cast posts. This study was conducted to measure the degree of cast posts taper, and to evaluate its suitability based on the anatomy aspects of the common candidate teeth for post reconstruction.</p> <p>Methods</p> <p>Working casts for cast posts, prepared using Gates Glidden drills, were collected. Impressions of post spaces were made using polyvinyl siloxan putty/wash technique. Digital camera with a 10' high quality lens was used for capturing two digital images for each impression; one in the Facio-Lingual (FL) and the other in the Mesio-Distal (MD) directions. Automated image processing program was developed to measure the degree of canal taper. Data were analyzed using Statistical Package for Social Sciences software and One way Analysis of Variance.</p> <p>Results</p> <p>Eighty four dies for cast posts were collected: 16 for each maxillary anterior teeth subgroup, and 18 for each maxillary and mandibular premolar subgroup. Mean of total taper for all preparations was 10.7 degree. There were no statistical differences among the total taper of all groups (P = .256) or between the MD and FL taper for each subgroup. Mean FL taper for the maxillary first premolars was lower significantly (P = .003) than the maxillary FL taper of the second premolars. FL taper was higher than the MD taper in all teeth except the maxillary first premolars.</p> <p>Conclusions</p> <p>Taper produced did not reflect the differences among the anatomy of teeth. While this technique deemed satisfactory in the maxillary anterior teeth, the same could not be said for the maxillary first premolars. Careful attention to the root anatomy is mandatory.</p
    corecore