249,049 research outputs found
Extending OpenVX for Model-based Design of Embedded Vision Applications
Developing computer vision applications for lowpower heterogeneous systems is increasingly gaining interest in the embedded systems community. Even more interesting is the tuning of such embedded software for the target architecture when this is driven by multiple constraints (e.g., performance, peak power, energy consumption). Indeed, developers frequently run into system-level inefficiencies and bottlenecks that can not be quickly addressed by traditional methods. In this context OpenVX has been proposed as the standard platform to develop portable, optimized and powerefficient applications for vision algorithms targeting embedded systems. Nevertheless, adopting OpenVX for rapid prototyping, early algorithm parametrization and validation of complex embedded applications is a very challenging task. This paper presents a methodology to integrate a model-based design environment to OpenVX. The methodology allows applying Matlab/Simulink for the model-based design, parametrization, and validation of computer vision applications. Then, it allows for the automatic synthesis of the application model into an OpenVX description for the hardware and constraints-aware application tuning. Experimental results have been conducted with an application for digital image stabilization developed through Simulink and, then, automatically synthesized into OpenVX-VisionWorks code for an NVIDIA Jetson TX1 boar
GPS driven camera selection in cyclocross races for automatic rider story generation
Cyclocross races are a very popular winter sport in Belgium and the Netherlands. In this paper we present a methodology to calculate the proximity of riders to a number of cameras that are located on a cyclocross course in order to automatically select the correct camera for each rider. The methodology is based on two main input sources. The first input is the course with cameras positioned along it. As the course and camera information is usually available as pdf and isn’t directly processable by computer programs, we propose the conversion GeoJSON. The second requirement for our methodology is accurate location tracking of the athletes on the course with the help of wearable GPS trackers. We present an experimental camera proximity algorithm that uses both input sources and finds for every rider at any given moment in the race the closest camera or vice versa. The output of this methodology results in automatic identification of the filmed riders by a given camera at a given moment in the race and might benefit post-processing of the camera video streams for further computer vision-based analysis of the streams, for example, to pre-filter the camera streams or to generate rider and team stories
A semantic-based probabilistic approach for real-time video event recognition
This is the author’s version of a work that was accepted for publication in Journal
Computer Vision and Image Understanding. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Journal
Computer Vision and Image Understanding, 116, 9 (2012) DOI: 10.1016/j.cviu.2012.04.005This paper presents an approach for real-time video event recognition that combines the accuracy and descriptive capabilities of, respectively, probabilistic and semantic approaches. Based on a state-of-art knowledge representation, we define a methodology for building recognition strategies from event descriptions that consider the uncertainty of the low-level analysis. Then, we efficiently organize such strategies for performing the recognition according to the temporal characteristics of events. In particular, we use Bayesian Networks and probabilistically-extended Petri Nets for recognizing, respectively, simple and complex events. For demonstrating the proposed approach, a framework has been implemented for recognizing human-object interactions in the video monitoring domain. The experimental results show that our approach improves the event recognition performance as compared to the widely used deterministic approach.This work has been partially supported by the Spanish Administration agency CDTI (CENIT-VISION 2007-
1007), by the Spanish Government (TEC2011-25995 EventVideo), by the ConsejerÃa de Educación of the Comunidad
de Madrid and by The European Social Fund
EyeRIS: A General-Purpose System for Eye Movement Contingent Display Control
In experimental studies of visual performance, the need often emerges to modify the stimulus according to the eye movements perfonncd by the subject. The methodology of Eye Movement-Contingent Display (EMCD) enables accurate control of the position and motion of the stimulus on the retina. EMCD procedures have been used successfully in many areas of vision science, including studies of visual attention, eye movements, and physiological characterization of neuronal response properties. Unfortunately, the difficulty of real-time programming and the unavailability of flexible and economical systems that can be easily adapted to the diversity of experimental needs and laboratory setups have prevented the widespread use of EMCD control. This paper describes EyeRIS, a general-purpose system for performing EMCD experiments on a Windows computer. Based on a digital signal processor with analog and digital interfaces, this integrated hardware and software system is responsible for sampling and processing oculomotor signals and subject responses and modifying the stimulus displayed on a CRT according to the gaze-contingent procedure specified by the experimenter. EyeRIS is designed to update the stimulus within a delay of 10 ms. To thoroughly evaluate EyeRIS' perforltlancc, this study (a) examines the response of the system in a number of EMCD procedures and computational benchmarking tests, (b) compares the accuracy of implementation of one particular EMCD procedure, retinal stabilization, to that produced by a standard tool used for this task, and (c) examines EyeRIS' performance in one of the many EMCD procedures that cannot be executed by means of any other currently available device.National Institute of Health (EY15732-01
Turning software engineers into machine learning engineers
A first challenge in teaching machine learning to software engineering and computer science students consists of changing the methodology from a constructive design-first perspective to an empirical one, focusing on proper experimental work. On the other hand, students nowadays can make significant progress using existing scripts and powerful (deep) learning frameworks -- focusing on established use cases such as vision tasks. To tackle problems in novel application domains, a clean methodological style is indispensable. Additionally, for deep learning, familiarity with gradient dynamics is crucial to understand deeper models. Consequently, we present three exercises that build upon each other to achieve these goals. These exercises are validated experimentally in a master's level course for software engineers
Integrating Simulink, OpenVX, and ROS for Model-Based Design of Embedded Vision Applications
OpenVX is increasingly gaining consensus as standard platform to develop portable, optimized and power-efficient embedded vision applications. Nevertheless, adopting OpenVX for rapid prototyping, early algorithm parametrization and validation of complex embedded applications is a very challenging task. This paper presents a comprehensive framework that integrates Simulink, OpenVX, and ROS for model-based design of embedded vision applications. The framework allows applying Matlab-Simulink for the model-based design, parametrization, and validation of computer vision applications. Then, it allows for the automatic synthesis of the application model into an OpenVX description for the hardware and constraints-aware application tuning. Finally, the methodology allows integrating the OpenVX application with Robot Operating System (ROS), which is the de-facto reference standard for developing robotic software applications. The OpenVX-ROS interface allows co-simulating and parametrizing the application by considering the actual robotic environment and the application reuse in any ROS-compliant system. Experimental results have been conducted with two real case studies: An application for digital image stabilization and the ORB descriptor for simultaneous localization and mapping (SLAM), which have been developed through Simulink and, then, automatically synthesized into OpenVX-VisionWorks code for an NVIDIA Jetson TX2 boar
Single-epoch supernova classification with deep convolutional neural networks
Supernovae Type-Ia (SNeIa) play a significant role in exploring the history
of the expansion of the Universe, since they are the best-known standard
candles with which we can accurately measure the distance to the objects.
Finding large samples of SNeIa and investigating their detailed characteristics
have become an important issue in cosmology and astronomy. Existing methods
relied on a photometric approach that first measures the luminance of supernova
candidates precisely and then fits the results to a parametric function of
temporal changes in luminance. However, it inevitably requires multi-epoch
observations and complex luminance measurements. In this work, we present a
novel method for classifying SNeIa simply from single-epoch observation images
without any complex measurements, by effectively integrating the
state-of-the-art computer vision methodology into the standard photometric
approach. Our method first builds a convolutional neural network for estimating
the luminance of supernovae from telescope images, and then constructs another
neural network for the classification, where the estimated luminance and
observation dates are used as features for classification. Both of the neural
networks are integrated into a single deep neural network to classify SNeIa
directly from observation images. Experimental results show the effectiveness
of the proposed method and reveal classification performance comparable to
existing photometric methods with multi-epoch observations.Comment: 7 pages, published as a workshop paper in ICDCS2017, in June 201
Relation-Oriented: Toward Knowledge-Aligned Causal AI
In machine learning, we naturally apply an Observation-Oriented principle, in
which observational variables preexist and set the stage for constructing
relationships. While sufficient for traditional models, the integration of AI
with big data exposes the misalignment between the observational models and our
actual comprehension. Contrarily, humans shape cognitive entities defined by
relationships, enabling us to formulate knowledge across temporal and
hyper-dimensional spaces, rather than being confined to observational
constructs. From an innovative Relation-Oriented perspective, this study
examines the roots of this misalignment within our current modeling paradigm,
illuminated by intuitive examples from computer vision and health informatics.
We also introduce the relation-defined representation learning methodology as a
practical implementation of Relation-Oriented modeling, supported by extensive
experimental validation
- …