139 research outputs found
Visual Tracking: An Experimental Survey
There is a large variety of trackers, which have been proposed in the literature during the last two decades with some mixed success. Object tracking in realistic scenarios is difficult problem, therefore it remains a most active area of research in Computer Vision. A good tracker should perform well in a large number of videos involving illumination changes, occlusion, clutter, camera motion, low contrast, specularities and at least six more aspects. However, the performance of proposed trackers have been evaluated typically on less than ten videos, or on the special purpose datasets. In this paper, we aim to evaluate trackers systematically and experimentally on 315 video fragments covering above aspects. We selected a set of nineteen trackers to include a wide variety of algorithms often cited in literature, supplemented with trackers appearing in 2010 and 2011 for which the code was publicly available. We demonstrate that trackers can be evaluated objectively by survival curves, Kaplan Meier statistics, and Grubs testing. We find that in the evaluation practice the F-score is as effective as the object tracking accuracy (OTA) score. The analysis under a large variety of circumstances provides objective insight into the strengths and weaknesses of trackers
Learning the Quality of Machine Permutations in Job Shop Scheduling
In recent years, the power demonstrated by Machine Learning (ML) has increasingly attracted the interest of the optimization community that is starting to leverage ML for enhancing and automating the design of algorithms. One combinatorial optimization problem recently tackled with ML is the Job Shop scheduling Problem (JSP). Most of the works on the JSP using ML focus on Deep Reinforcement Learning (DRL), and only a few of them leverage supervised learning techniques. The recurrent reasons for avoiding supervised learning seem to be the difficulty in casting the right learning task, i.e., what is meaningful to predict, and how to obtain labels. Therefore, we first propose a novel supervised learning task that aims at predicting the quality of machine permutations. Then, we design an original methodology to estimate this quality, and we use these estimations to create an accurate sequential deep learning model (binary accuracy above 95%). Finally, we empirically demonstrate the value of predicting the quality of machine permutations by enhancing the performance of a simple Tabu Search algorithm inspired by the works in the literature
Quick, accurate, smart: 3D computer vision technology helps assessing confined animals' behaviour
Mankind directly controls the environment and lifestyles of several domestic species for purposes ranging from production and research to conservation and companionship. These environments and lifestyles may not offer these animals the best quality of life. Behaviour is a direct reflection of how the animal is coping with its environment. Behavioural indicators are thus among the preferred parameters to assess welfare. However, behavioural recording (usually from video) can be very time consuming and the accuracy and reliability of the output rely on the experience and background of the observers. The outburst of new video technology and computer image processing gives the basis for promising solutions. In this pilot study, we present a new prototype software able to automatically infer the behaviour of dogs housed in kennels from 3D visual data and through structured machine learning frameworks. Depth information acquired through 3D features, body part detection and training are the key elements that allow the machine to recognise postures, trajectories inside the kennel and patterns of movement that can be later labelled at convenience. The main innovation of the software is its ability to automatically cluster frequently observed temporal patterns of movement without any pre-set ethogram. Conversely, when common patterns are defined through training, a deviation from normal behaviour in time or between individuals could be assessed. The software accuracy in correctly detecting the dogs' behaviour was checked through a validation process. An automatic behaviour recognition system, independent from human subjectivity, could add scientific knowledge on animals' quality of life in confinement as well as saving time and resources. This 3D framework was designed to be invariant to the dog's shape and size and could be extended to farm, laboratory and zoo quadrupeds in artificial housing. The computer vision technique applied to this software is innovative in non-human animal behaviour science. Further improvements and validation are needed, and future applications and limitations are discussed.</p
Hypomineralized Second Primary Molars as Predictor of Molar Incisor Hypomineralization
Molar incisor hypomineralization (MIH) is a developmental defect of dental enamel that shares features with hypomineralized second primary molars (HSPM). Prior to permanent tooth eruption, second primary molars could have predictive value for permanent molar and incisor hypomineralization. To assess this possible relationship, a cross-sectional study was conducted in a sample of 414 children aged 8 and 9 years from the INMA cohort in Valencia (Spain). A calibrated examiner (linear-weighted Kappa 0.83) performed the intraoral examinations at the University of Valencia between November 2013 and 2014, applying the diagnostic criteria for MIH and HSPM adopted by the European Academy of Paediatric Dentistry. 100 children (24.2%) presented MIH and 60 (14.5%) presented HSPM. Cooccurrence of the two defects was observed in 11.1% of the children examined. The positive predictive value was 76.7% (63.9-86.6) and the negative predictive value 84.7% (80.6-88.3). The positive likelihood ratio (S/1-E) was 10.3 (5.9-17.9) and the negative likelihood ratio (1-S/E) 0.57 (0.47-0.68). The odds ratio was 18.2 (9.39-35.48). It was concluded that while the presence of HSPM can be considered a predictor of MIH, indicating the need for monitoring and control, the absence of this defect in primary dentition does not rule out the appearance of MIH
Thrombin generation in a woman with heterozygous factor V Leiden and combined oral contraceptives: A case report.
Combined oral contraceptives and factor V Leiden mutation are multiplicative risk factors for venous thromboembolism. However, it remains unknown whether this multiplicative effect is reflected in thrombin generation assays. We report here the evolution of the thrombin generation profile while taking combined oral contraceptives and after their discontinuation in a woman with heterozygous factor V Leiden mutation. The proband exhibited a distinctly prothrombotic thrombin generation profile including markedly decreased thrombomodulin (TM) sensitivity, compared to the control population. This profile possibly reflected a high thrombotic risk. After discontinuation of combined oral contraceptives, thrombin generation and TM sensitivity improved greatly, leaving only a slightly prothrombotic profile. Therefore, the multiplied thrombotic risk occurring with simultaneous combined oral contraceptives and factor V Leiden mutation is reflected by a thrombin generation assay performed without and with TM. This could be a promising tool to identify women taking combined oral contraceptives at high risk for venous thromboembolism. Further studies are needed to verify this hypothesis
ClusterFix: A Cluster-Based Debiasing Approach without Protected-Group Supervision
The failures of Deep Networks can sometimes be ascribed to biases in the data or algorithmic choices. Existing debiasing approaches exploit prior knowledge to avoid unintended solutions; we acknowledge that, in real-world settings, it could be unfeasible to gather enough prior information to characterize the bias, or it could even raise ethical considerations. We hence propose a novel debiasing approach, termed ClusterFix, which does not require any external hint about the nature of biases. Such an approach alters the standard empirical risk minimization and introduces a per-example weight, encoding how critical and far from the majority an example is. Notably, the weights consider how difficult it is for the model to infer the correct pseudo-label, which is obtained in a self-supervised manner by dividing examples into multiple clusters. Extensive experiments show that the misclassification error incurred in identifying the correct cluster allows for identifying examples prone to bias-related issues. As a result, our approach outperforms existing methods on standard benchmarks for bias removal and fairness
How many Observations are Enough? Knowledge Distillation for Trajectory Forecasting
Accurate prediction of future human positions is an essential task for modern video-surveillance systems. Current state-of-the-art models usually rely on a "history" of past tracked locations (e.g., 3 to 5 seconds) to predict a plausible sequence of future locations (e.g., up to the next 5 seconds). We feel that this common schema neglects critical traits of realistic applications: as the collection of input trajectories involves machine perception (i.e., detection and tracking), incorrect detection and fragmentation errors may accumulate in crowded scenes, leading to tracking drifts. On this account, the model would be fed with corrupted and noisy input data, thus fatally affecting its prediction performance.In this regard, we focus on delivering accurate predictions when only few input observations are used, thus potentially lowering the risks associated with automatic perception. To this end, we conceive a novel distillation strategy that allows a knowledge transfer from a teacher network to a student one, the latter fed with fewer observations (just two ones). We show that a properly defined teacher supervision allows a student network to perform comparably to state-of-the-art approaches that demand more observations. Besides, extensive experiments on common trajectory forecasting datasets highlight that our student network better generalizes to unseen scenarios
Input Perturbation Reduces Exposure Bias in Diffusion Models
Denoising Diffusion Probabilistic Models have shown an impressive generation quality although their long sampling chain leads to high computational costs. In this paper, we observe that a long sampling chain also leads to an error accumulation phenomenon, which is similar to the exposure bias problem in autoregressive text generation. Specifically, we note that there is a discrepancy between training and testing, since the former is conditioned on the ground truth samples, while the latter is conditioned on the previously generated results. To alleviate this problem, we propose a very simple but effective training regularization, consisting in perturbing the ground truth samples to simulate the inference time prediction errors. We empirically show that, without affecting the recall and precision, the proposed input perturbation leads to a significant improvement in the sample quality while reducing both the training and the inference times. For instance, on CelebA 64×64, we achieve a new state-of-the-art FID score of 1.27, while saving 37.5% of the training time. The code is available at https://github.com/forever208/DDPM-IP
AC-VRNN: Attentive Conditional-VRNN for multi-future trajectory prediction
Anticipating human motion in crowded scenarios is essential for developing intelligent transportation systems, social-aware robots and advanced video surveillance applications. A key component of this task is represented by the inherently multi-modal nature of human paths which makes socially acceptable multiple futures when human interactions are involved. To this end, we propose a generative architecture for multi-future trajectory predictions based on Conditional Variational Recurrent Neural Networks (C-VRNNs). Conditioning mainly relies on prior belief maps, representing most likely moving directions and forcing the model to consider past observed dynamics in generating future positions. Human interactions are modelled with a graph-based attention mechanism enabling an online attentive hidden state refinement of the recurrent estimation. To corroborate our model, we perform extensive experiments on publicly-available datasets (e.g., ETH/UCY, Stanford Drone Dataset, STATS SportVU NBA, Intersection Drone Dataset and TrajNet++) and demonstrate its effectiveness in crowded scenes compared to several state-of-the-art methods
Thrombocytopathies: Not Just Aggregation Defects-The Clinical Relevance of Procoagulant Platelets.
Platelets are active key players in haemostasis. Qualitative platelet dysfunctions result in thrombocytopathies variously characterized by defects of their adhesive and procoagulant activation endpoints. In this review, we summarize the traditional platelet defects in adhesion, secretion, and aggregation. In addition, we review the current knowledge about procoagulant platelets, focusing on their role in bleeding or thrombotic pathologies and their pharmaceutical modulation. Procoagulant activity is an important feature of platelet activation, which should be specifically evaluated during the investigation of a suspected thrombocytopathy
- …
