67 research outputs found

    Variational Clustering: Leveraging Variational Autoencoders for Image Clustering

    Full text link
    Recent advances in deep learning have shown their ability to learn strong feature representations for images. The task of image clustering naturally requires good feature representations to capture the distribution of the data and subsequently differentiate data points from one another. Often these two aspects are dealt with independently and thus traditional feature learning alone does not suffice in partitioning the data meaningfully. Variational Autoencoders (VAEs) naturally lend themselves to learning data distributions in a latent space. Since we wish to efficiently discriminate between different clusters in the data, we propose a method based on VAEs where we use a Gaussian Mixture prior to help cluster the images accurately. We jointly learn the parameters of both the prior and the posterior distributions. Our method represents a true Gaussian Mixture VAE. This way, our method simultaneously learns a prior that captures the latent distribution of the images and a posterior to help discriminate well between data points. We also propose a novel reparametrization of the latent space consisting of a mixture of discrete and continuous variables. One key takeaway is that our method generalizes better across different datasets without using any pre-training or learnt models, unlike existing methods, allowing it to be trained from scratch in an end-to-end manner. We verify our efficacy and generalizability experimentally by achieving state-of-the-art results among unsupervised methods on a variety of datasets. To the best of our knowledge, we are the first to pursue image clustering using VAEs in a purely unsupervised manner on real image datasets

    Human-Robot Handshaking: A Review

    Full text link
    For some years now, the use of social, anthropomorphic robots in various situations has been on the rise. These are robots developed to interact with humans and are equipped with corresponding extremities. They already support human users in various industries, such as retail, gastronomy, hotels, education and healthcare. During such Human-Robot Interaction (HRI) scenarios, physical touch plays a central role in the various applications of social robots as interactive non-verbal behaviour is a key factor in making the interaction more natural. Shaking hands is a simple, natural interaction used commonly in many social contexts and is seen as a symbol of greeting, farewell and congratulations. In this paper, we take a look at the existing state of Human-Robot Handshaking research, categorise the works based on their focus areas, draw out the major findings of these areas while analysing their pitfalls. We mainly see that some form of synchronisation exists during the different phases of the interaction. In addition to this, we also find that additional factors like gaze, voice facial expressions etc. can affect the perception of a robotic handshake and that internal factors like personality and mood can affect the way in which handshaking behaviours are executed by humans. Based on the findings and insights, we finally discuss possible ways forward for research on such physically interactive behaviours.Comment: Pre-print version. Accepted for publication in the International Journal of Social Robotic

    Significance of Diabetes (Siphon) in General And Orthodontic Treatment – A Literature Review

    Get PDF
    Diabetes Mellitus is a metabolic disorder of carbohydrate, protein and fat resulting from defective synthesis of insulin or its action on body tissue. It is one of the most commonly affected diseases worldwide and India has increasing prevalence in 2018 from 45% to 64%. Some of the deleterious effects of diabetes on oral health include xerostomia, debris accumulation,dental caries, recurrent infections, periodontitis etc. Periodontal destruction is the limiting factor in the orthodontic treatment. Maintaining oral hygiene and prevention of periodontal destruction is important before seeking for orthodontic treatment. Many researches and advancements made for periodontally compromised patients which include low profilebrackets, copper NiTi wires and additional anchorages like mini screws. This article emphasizes on significance of diabetes in general and in orthodontic patients and treatment modalities of such patients

    Learning to Prevent Monocular SLAM Failure using Reinforcement Learning

    Full text link
    Monocular SLAM refers to using a single camera to estimate robot ego motion while building a map of the environment. While Monocular SLAM is a well studied problem, automating Monocular SLAM by integrating it with trajectory planning frameworks is particularly challenging. This paper presents a novel formulation based on Reinforcement Learning (RL) that generates fail safe trajectories wherein the SLAM generated outputs do not deviate largely from their true values. Quintessentially, the RL framework successfully learns the otherwise complex relation between perceptual inputs and motor actions and uses this knowledge to generate trajectories that do not cause failure of SLAM. We show systematically in simulations how the quality of the SLAM dramatically improves when trajectories are computed using RL. Our method scales effectively across Monocular SLAM frameworks in both simulation and in real world experiments with a mobile robot.Comment: Accepted at the 11th Indian Conference on Computer Vision, Graphics and Image Processing (ICVGIP) 2018 More info can be found at the project page at https://robotics.iiit.ac.in/people/vignesh.prasad/SLAMSafePlanner.html and the supplementary video can be found at https://www.youtube.com/watch?v=420QmM_Z8v

    Detection rates of recurrent prostate cancer : 68Gallium (Ga)-labelled prostate-specific membrane antigen versus choline PET/CT scans. A systematic review

    Get PDF
    Background: The aim of this work was to assess the use of prostate-specific membrane antigen (PSMA)-labelled radiotracers in detecting the recurrence of prostate cancer. PSMA is thought to have higher detection rates when utilized in positron emission tomography (PET)/computed tomography (CT) scans, particularly at lower prostate-specific antigen (PSA) levels, compared with choline-based scans. Methods: A systematic review was conducted comparing choline and PSMA PET/CT scans in patients with recurrent prostate cancer following an initial curative attempt. The primary outcomes were overall detection rates, detection rates at low PSA thresholds, difference in detection rates and exclusive detection rates on a per-person analysis. Secondary outcome measures were total number of lesions, exclusive detection by each scan on a per-lesion basis and adverse side effects. Results: Overall detection rates were 79.8% for PSMA and 66.7% for choline. There was a statistically significant difference in detection rates favouring PSMA [OR (M–H, random, 95% confidence interval (CI)) 2.27 (1.06, 4.85), p = 0.04]. Direct comparison was limited to PSA < 2 ng/ml in two studies, with no statistically significant difference in detection rates between the scans [OR (M–H, random, 95% CI) 2.37 (0.61, 9.17) p = 0.21]. The difference in detection on the per-patient analysis was significantly higher in the PSMA scans (p < 0.00001). All three studies reported higher lymph node, bone metastasis and locoregional recurrence rates in PSMA. Conclusions: PSMA PET/CT has a better performance compared with choline PET/CT in detecting recurrent disease both on per-patient and per-lesion analysis and should be the imaging modality of choice while deciding on salvage and nonsystematic metastasis-directed therapy strategies.Peer reviewedFinal Published versio

    MILD: Multimodal Interactive Latent Dynamics for Learning Human-Robot Interaction

    Full text link
    Modeling interaction dynamics to generate robot trajectories that enable a robot to adapt and react to a human's actions and intentions is critical for efficient and effective collaborative Human-Robot Interactions (HRI). Learning from Demonstration (LfD) methods from Human-Human Interactions (HHI) have shown promising results, especially when coupled with representation learning techniques. However, such methods for learning HRI either do not scale well to high dimensional data or cannot accurately adapt to changing via-poses of the interacting partner. We propose Multimodal Interactive Latent Dynamics (MILD), a method that couples deep representation learning and probabilistic machine learning to address the problem of two-party physical HRIs. We learn the interaction dynamics from demonstrations, using Hidden Semi-Markov Models (HSMMs) to model the joint distribution of the interacting agents in the latent space of a Variational Autoencoder (VAE). Our experimental evaluations for learning HRI from HHI demonstrations show that MILD effectively captures the multimodality in the latent representations of HRI tasks, allowing us to decode the varying dynamics occurring in such tasks. Compared to related work, MILD generates more accurate trajectories for the controlled agent (robot) when conditioned on the observed agent's (human) trajectory. Notably, MILD can learn directly from camera-based pose estimations to generate trajectories, which we then map to a humanoid robot without the need for any additional training.Comment: Accepted at the IEEE-RAS International Conference on Humanoid Robots (Humanoids) 202

    Evaluation of the Handshake Turing Test for anthropomorphic Robots

    Full text link
    Handshakes are fundamental and common greeting and parting gestures among humans. They are important in shaping first impressions as people tend to associate character traits with a person's handshake. To widen the social acceptability of robots and make a lasting first impression, a good handshaking ability is an important skill for social robots. Therefore, to test the human-likeness of a robot handshake, we propose an initial Turing-like test, primarily for the hardware interface to future AI agents. We evaluate the test on an android robot's hand to determine if it can pass for a human hand. This is an important aspect of Turing tests for motor intelligence where humans have to interact with a physical device rather than a virtual one. We also propose some modifications to the definition of a Turing test for such scenarios taking into account that a human needs to interact with a physical medium.Comment: Accepted as a Late Breaking Report in The 15th Annual ACM/IEEE International Conference on Human Robot Interaction (HRI) 202
    • …
    corecore