41 research outputs found

    Congestion Control for Network-Aware Telehaptic Communication

    Full text link
    Telehaptic applications involve delay-sensitive multimedia communication between remote locations with distinct Quality of Service (QoS) requirements for different media components. These QoS constraints pose a variety of challenges, especially when the communication occurs over a shared network, with unknown and time-varying cross-traffic. In this work, we propose a transport layer congestion control protocol for telehaptic applications operating over shared networks, termed as dynamic packetization module (DPM). DPM is a lossless, network-aware protocol which tunes the telehaptic packetization rate based on the level of congestion in the network. To monitor the network congestion, we devise a novel network feedback module, which communicates the end-to-end delays encountered by the telehaptic packets to the respective transmitters with negligible overhead. Via extensive simulations, we show that DPM meets the QoS requirements of telehaptic applications over a wide range of network cross-traffic conditions. We also report qualitative results of a real-time telepottery experiment with several human subjects, which reveal that DPM preserves the quality of telehaptic activity even under heavily congested network scenarios. Finally, we compare the performance of DPM with several previously proposed telehaptic communication protocols and demonstrate that DPM outperforms these protocols.Comment: 25 pages, 19 figure

    A Study of Velocity-Dependent JND of Haptic Model Detail

    Get PDF
    The study of haptics, or the sense of touch in virtual reality environments, is constantly looking for improvements in modeling with a high fidelity. Highly detailed models are desirable, but they often lead to slow processing times, which can mean a loss of fidelity in the force feedback sensations. Model compression techniques are critical to balancing model detail and processing time. One of the proposed compression techniques is to create multiple models of the same object but with different levels of detail (LOD) for each model. The technique hypothesizes that the human arm loses sensitivity to forces with the increase of its movement speed. This the compression technique determines which model to use based on the user's movement speed. This dissertation examines studies how the movement speed of the user affects the user's ability to sense changes in details of haptic models. Experiments are conducted using different haptic surfaces. Their levels of detail are changed while the subject interacts with them to mimic the effects of a multiresolution compression implementation. The tests focus on the subjects' ability to differentiate changes of the surfaces at each speed. The first experiment uses curved surfaces with multiple resolutions. This test observes the sensitivity of the user when the details on the surface are small. The results show that the subjects are more sensitive to changes of small details at a lower speed than higher speed. The second experiment measures sensitivity to larger features by using trapezoidal surfaces with different angles. The trapezoidal surfaces can be seen as a low-resolution haptic model with only two vertices, and changing the angles of the trapezoids is seen as changing the radii of curvature. With the same speed settings from the first experiment applied to the subjects, the sensitivity for changes in curvature is predicted to decrease with the increase of speed. However, the results of this experiment proved otherwise. The conclusions suggest that multiresolution designs are not a straightforward reduction of LOD, even though the movement speed does affect haptic sensitivity. The model's geometry should be taken into account when designing the parameters for haptic model compression. The results from the experiments provide insights to future haptic multiresolution compression designs

    Factors Affecting Human Force Perception and Performance in Haptic-Enabled Virtual Environments

    Get PDF
    Haptic technology enables computer users to touch and/or manipulate virtual objects in virtual environments (VEs). Similar to other human-in-the-loop applications, haptic applications require interactions between humans and computers. Thus, human-factors studies are required to recognize the limitations and capabilities of the user. This thesis establishes human-factors criteria to improve various haptic applications such as perception-based haptic compression techniques and haptic-enabled computer-aided design (CAD). Today, data compression plays a significant role in the transmission of haptic information since the efficient use of the available bandwidth is a concern. Most lossy haptic compression techniques rely on the limitations of human force perception, and this is used in the design of perception-based haptic compression techniques. Researchers have studied force perception when a user is in static interaction with a stationary object. This thesis focuses on cases where the human user and the object are in relative motion. The limitations of force perception are quantified using psychophysical methods, and the effects of several factors, including user hand velocity and sensory adaptation, are investigated. The results indicate that fewer haptic details need to be calculated or transmitted when the user's hand is in motion. In traditional CAD systems, users usually design virtual prototypes using a mouse via their vision system only, and it is difficult to design curved surfaces due to the number, shape, and position of the curves. Adding haptics to CAD systems enables users to explore and manipulate virtual objects using the sense of touch. In addition, human performance is important in CAD environments. To maintain the accuracy, active haptic manipulation of the user response can be incorporated in CAD applications. This thesis investigates the effect of forces on the accuracy of movement in VEs. The results indicate that factors such as the base force intensity and force increment/decrement can be incorporated in the control of users' movements in VEs. In other words, we can pull/push the users' hands by increasing/decreasing the force without the users being aware of it

    Factors Affecting Human Force Perception and Performance in Haptic-Enabled Virtual Environments

    Get PDF
    Haptic technology enables computer users to touch and/or manipulate virtual objects in virtual environments (VEs). Similar to other human-in-the-loop applications, haptic applications require interactions between humans and computers. Thus, human-factors studies are required to recognize the limitations and capabilities of the user. This thesis establishes human-factors criteria to improve various haptic applications such as perception-based haptic compression techniques and haptic-enabled computer-aided design (CAD). Today, data compression plays a significant role in the transmission of haptic information since the efficient use of the available bandwidth is a concern. Most lossy haptic compression techniques rely on the limitations of human force perception, and this is used in the design of perception-based haptic compression techniques. Researchers have studied force perception when a user is in static interaction with a stationary object. This thesis focuses on cases where the human user and the object are in relative motion. The limitations of force perception are quantified using psychophysical methods, and the effects of several factors, including user hand velocity and sensory adaptation, are investigated. The results indicate that fewer haptic details need to be calculated or transmitted when the user's hand is in motion. In traditional CAD systems, users usually design virtual prototypes using a mouse via their vision system only, and it is difficult to design curved surfaces due to the number, shape, and position of the curves. Adding haptics to CAD systems enables users to explore and manipulate virtual objects using the sense of touch. In addition, human performance is important in CAD environments. To maintain the accuracy, active haptic manipulation of the user response can be incorporated in CAD applications. This thesis investigates the effect of forces on the accuracy of movement in VEs. The results indicate that factors such as the base force intensity and force increment/decrement can be incorporated in the control of users' movements in VEs. In other words, we can pull/push the users' hands by increasing/decreasing the force without the users being aware of it

    Mulsemedia: State of the art, perspectives, and challenges

    Get PDF
    Mulsemedia-multiple sensorial media-captures a wide variety of research efforts and applications. This article presents a historic perspective on mulsemedia work and reviews current developments in the area. These take place across the traditional multimedia spectrum-from virtual reality applications to computer games-as well as efforts in the arts, gastronomy, and therapy, to mention a few. We also describe standardization efforts, via the MPEG-V standard, and identify future developments and exciting challenges the community needs to overcome

    Haptic data reduction through dynamic perceptual analysis and event-based communication

    Full text link
    This research presents an adjustable and flexible framework for haptic data compression and communication that can be used in a robotic teleoperation session. The framework contains a customized event-driven transmission control protocol, several dynamically adaptive perceptual and prediction methods for haptic sample reduction, and last but not the least, an architecture for the data flow

    Modelling the human perception of shape-from-shading

    Get PDF
    Shading conveys information on 3-D shape and the process of recovering this information is called shape-from-shading (SFS). This thesis divides the process of human SFS into two functional sub-units (luminance disambiguation and shape computation) and studies them individually. Based on results of a series of psychophysical experiments it is proposed that the interaction between first- and second-order channels plays an important role in disambiguating luminance. Based on this idea, two versions of a biologically plausible model are developed to explain the human performances observed here and elsewhere. An algorithm sharing the same idea is also developed as a solution to the problem of intrinsic image decomposition in the field of image processing. With regard to the shape computation unit, a link between luminance variations and estimated surface norms is identified by testing participants on simple gratings with several different luminance profiles. This methodology is unconventional but can be justified in the light of past studies of human SFS. Finally a computational algorithm for SFS containing two distinct operating modes is proposed. This algorithm is broadly consistent with the known psychophysics on human SFS

    Change blindness: eradication of gestalt strategies

    Get PDF
    Arrays of eight, texture-defined rectangles were used as stimuli in a one-shot change blindness (CB) task where there was a 50% chance that one rectangle would change orientation between two successive presentations separated by an interval. CB was eliminated by cueing the target rectangle in the first stimulus, reduced by cueing in the interval and unaffected by cueing in the second presentation. This supports the idea that a representation was formed that persisted through the interval before being 'overwritten' by the second presentation (Landman et al, 2003 Vision Research 43149–164]. Another possibility is that participants used some kind of grouping or Gestalt strategy. To test this we changed the spatial position of the rectangles in the second presentation by shifting them along imaginary spokes (by ±1 degree) emanating from the central fixation point. There was no significant difference seen in performance between this and the standard task [F(1,4)=2.565, p=0.185]. This may suggest two things: (i) Gestalt grouping is not used as a strategy in these tasks, and (ii) it gives further weight to the argument that objects may be stored and retrieved from a pre-attentional store during this task
    corecore