288 research outputs found
Massively Parallel Video Networks
We introduce a class of causal video understanding models that aims to
improve efficiency of video processing by maximising throughput, minimising
latency, and reducing the number of clock cycles. Leveraging operation
pipelining and multi-rate clocks, these models perform a minimal amount of
computation (e.g. as few as four convolutional layers) for each frame per
timestep to produce an output. The models are still very deep, with dozens of
such operations being performed but in a pipelined fashion that enables
depth-parallel computation. We illustrate the proposed principles by applying
them to existing image architectures and analyse their behaviour on two video
tasks: action recognition and human keypoint localisation. The results show
that a significant degree of parallelism, and implicitly speedup, can be
achieved with little loss in performance.Comment: Fixed typos in densenet model definition in appendi
Self-Supervised Relative Depth Learning for Urban Scene Understanding
As an agent moves through the world, the apparent motion of scene elements is
(usually) inversely proportional to their depth. It is natural for a learning
agent to associate image patterns with the magnitude of their displacement over
time: as the agent moves, faraway mountains don't move much; nearby trees move
a lot. This natural relationship between the appearance of objects and their
motion is a rich source of information about the world. In this work, we start
by training a deep network, using fully automatic supervision, to predict
relative scene depth from single images. The relative depth training images are
automatically derived from simple videos of cars moving through a scene, using
recent motion segmentation techniques, and no human-provided labels. This proxy
task of predicting relative depth from a single image induces features in the
network that result in large improvements in a set of downstream tasks
including semantic segmentation, joint road segmentation and car detection, and
monocular (absolute) depth estimation, over a network trained from scratch. The
improvement on the semantic segmentation task is greater than those produced by
any other automatically supervised methods. Moreover, for monocular depth
estimation, our unsupervised pre-training method even outperforms supervised
pre-training with ImageNet. In addition, we demonstrate benefits from learning
to predict (unsupervised) relative depth in the specific videos associated with
various downstream tasks. We adapt to the specific scenes in those tasks in an
unsupervised manner to improve performance. In summary, for semantic
segmentation, we present state-of-the-art results among methods that do not use
supervised pre-training, and we even exceed the performance of supervised
ImageNet pre-trained models for monocular depth estimation, achieving results
that are comparable with state-of-the-art methods
An experimental multiprocessor system for distributed parallel computations.
The availability of low-cost microprocessor chips with efficient instruction sets for specific numerical tasks (signal processors) has been exploited for building a versatile multiprocessor system, consisting of a host minicomputer augmented by a number of joint processors. The host provides a multiuser-multitasking environment and manages system resources and task scheduling. User applications can call upon one or more joint processors for parallel execution of adequately partitioned, computationally intensive numeric operations. Each joint processor has sufficient local memory for storing procedures and data and has access to regions in host memory for shared data. Kernel processes in the host and in the joint processors provide the necessary mechanism for initialization and synchronization of the distributed parallel execution of procedures
Arsenic induces metabolome remodeling in mature human adipocytes.
Human lifetime exposure to arsenic through drinking water, food supply or industrial pollution leads to its accumulation in many organs such as liver, kidneys, lungs or pancreas but also adipose tissue. Recently, population-based studies revealed the association between arsenic exposure and the development of metabolic diseases such as obesity and type 2 diabetes. To shed light on the molecular bases of such association, we determined the concentration that inhibited 17% of cell viability and investigated the effects of arsenic acute exposure on adipose-derived human mesenchymal stem cells differentiated in vitro into mature adipocytes and treated with sodium arsenite (NaAsO <sub>2</sub> , 10 nM to 10 µM). Untargeted metabolomics and gene expression analyses revealed a strong dose-dependent inhibition of lipogenesis and lipolysis induction, reducing the cellular ability to store lipids. These dysregulations were emphasized by the inhibition of the cellular response to insulin, as shown by the perturbation of several genes and metabolites involved in the mentioned biological pathways. Our study highlighted the activation of an adaptive oxidative stress response with the strong induction of metallothioneins and increased glutathione levels in response to arsenic accumulation that could exacerbate the decreased insulin sensitivity of the adipocytes. Arsenic exposure strongly affected the expression of arsenic transporters, responsible for arsenic influx and efflux, and induced a pro-inflammatory state in adipocytes by enhancing the expression of the inflammatory interleukin 6 (IL6). Collectively, our data showed that an acute exposure to low levels of arsenic concentrations alters key adipocyte functions, highlighting its contribution to the development of insulin resistance and the pathogenesis of metabolic disorders
View-tolerant face recognition and Hebbian learning imply mirror-symmetric neural tuning to head orientation
The primate brain contains a hierarchy of visual areas, dubbed the ventral
stream, which rapidly computes object representations that are both specific
for object identity and relatively robust against identity-preserving
transformations like depth-rotations. Current computational models of object
recognition, including recent deep learning networks, generate these properties
through a hierarchy of alternating selectivity-increasing filtering and
tolerance-increasing pooling operations, similar to simple-complex cells
operations. While simulations of these models recapitulate the ventral stream's
progression from early view-specific to late view-tolerant representations,
they fail to generate the most salient property of the intermediate
representation for faces found in the brain: mirror-symmetric tuning of the
neural population to head orientation. Here we prove that a class of
hierarchical architectures and a broad set of biologically plausible learning
rules can provide approximate invariance at the top level of the network. While
most of the learning rules do not yield mirror-symmetry in the mid-level
representations, we characterize a specific biologically-plausible Hebb-type
learning rule that is guaranteed to generate mirror-symmetric tuning to faces
tuning at intermediate levels of the architecture
A comparison of two computer-based face identification systems with human perceptions of faces
The performance of two different computer systems for representing faces was compared with human ratings of similarity and distinctiveness, and human memory performance, on a specific set of face images. The systems compared were a graphmatching system (e.g. Lades et al., 1993) and coding based on Principal Components Analysis (PCA) of image pixels (e.g. Turk & Pentland, 1991). Replicating other work, the PCA-based system produced very much better performance at recognising faces, and higher correlations with human performance with the same images, when the images were initially standardised using a morphing procedure and separate analysis of "shape" and "shape-free" components then combined. Both the graph-matching and (shape + shape-free) PCA systems were equally able to recognise faces shown with changed expressions, both provided reasonable correlations with human ratings and memory data, and there were also correlations between the facial similarities recorded by each of the computer models. However, comparisons with human similarity ratings of faces with and without the hair visible, and prediction of memory performance with and without alteration in face expressions, suggested that the graph-matching system was better at capturing aspects of the appearance of the face, while the PCA-based system seemed better at capturing aspects of the appearance of specific images of faces
Deep Learning of Representations: Looking Forward
Deep learning research aims at discovering learning algorithms that discover
multiple levels of distributed representations, with higher levels representing
more abstract concepts. Although the study of deep learning has already led to
impressive theoretical results, learning algorithms and breakthrough
experiments, several challenges lie ahead. This paper proposes to examine some
of these challenges, centering on the questions of scaling deep learning
algorithms to much larger models and datasets, reducing optimization
difficulties due to ill-conditioning or local minima, designing more efficient
and powerful inference and sampling procedures, and learning to disentangle the
factors of variation underlying the observed data. It also proposes a few
forward-looking research directions aimed at overcoming these challenges
The taper of cast post preparation measured using innovative image processing technique
<p>Abstract</p> <p>Background</p> <p>No documentation in the literature about taper of cast posts. This study was conducted to measure the degree of cast posts taper, and to evaluate its suitability based on the anatomy aspects of the common candidate teeth for post reconstruction.</p> <p>Methods</p> <p>Working casts for cast posts, prepared using Gates Glidden drills, were collected. Impressions of post spaces were made using polyvinyl siloxan putty/wash technique. Digital camera with a 10' high quality lens was used for capturing two digital images for each impression; one in the Facio-Lingual (FL) and the other in the Mesio-Distal (MD) directions. Automated image processing program was developed to measure the degree of canal taper. Data were analyzed using Statistical Package for Social Sciences software and One way Analysis of Variance.</p> <p>Results</p> <p>Eighty four dies for cast posts were collected: 16 for each maxillary anterior teeth subgroup, and 18 for each maxillary and mandibular premolar subgroup. Mean of total taper for all preparations was 10.7 degree. There were no statistical differences among the total taper of all groups (P = .256) or between the MD and FL taper for each subgroup. Mean FL taper for the maxillary first premolars was lower significantly (P = .003) than the maxillary FL taper of the second premolars. FL taper was higher than the MD taper in all teeth except the maxillary first premolars.</p> <p>Conclusions</p> <p>Taper produced did not reflect the differences among the anatomy of teeth. While this technique deemed satisfactory in the maxillary anterior teeth, the same could not be said for the maxillary first premolars. Careful attention to the root anatomy is mandatory.</p
- …