487 research outputs found
Haptic Interaction with 3D oriented point clouds on the GPU
Real-time point-based rendering and interaction with virtual objects is gaining popularity
and importance as di�erent haptic devices and technologies increasingly provide the basis
for realistic interaction. Haptic Interaction is being used for a wide range of applications
such as medical training, remote robot operators, tactile displays and video games. Virtual
object visualization and interaction using haptic devices is the main focus; this process
involves several steps such as: Data Acquisition, Graphic Rendering, Haptic Interaction
and Data Modi�cation. This work presents a framework for Haptic Interaction using the
GPU as a hardware accelerator, and includes an approach for enabling the modi�cation
of data during interaction. The results demonstrate the limits and capabilities of these
techniques in the context of volume rendering for haptic applications. Also, the use
of dynamic parallelism as a technique to scale the number of threads needed from the
accelerator according to the interaction requirements is studied allowing the editing of
data sets of up to one million points at interactive haptic frame rates
Painterly rendering techniques: A state-of-the-art review of current approaches
In this publication we will look at the different methods presented over the past few decades which attempt to recreate digital paintings. While previous surveys concentrate on the broader subject of non-photorealistic rendering, the focus of this paper is firmly placed on painterly rendering techniques. We compare different methods used to produce different output painting styles such as abstract, colour pencil, watercolour, oriental, oil and pastel. Whereas some methods demand a high level of interaction using a skilled artist, others require simple parameters provided by a user with little or no artistic experience. Many methods attempt to provide more automation with the use of varying forms of reference data. This reference data can range from still photographs, video, 3D polygonal meshes or even 3D point clouds. The techniques presented here endeavour to provide tools and styles that are not traditionally available to an artist. Copyright © 2012 John Wiley & Sons, Ltd
Volumetric cloud generation using a Chinese brush calligraphy style
Includes bibliographical references.Clouds are an important feature of any real or simulated environment in which the sky is visible. Their amorphous, ever-changing and illuminated features make the sky vivid and beautiful. However, these features increase both the complexity of real time rendering and modelling. It is difficult to design and build volumetric clouds in an easy and intuitive way, particularly if the interface is intended for artists rather than programmers. We propose a novel modelling system motivated by an ancient painting style, Chinese Landscape Painting, to address this problem. With the use of only one brush and one colour, an artist can paint a vivid and detailed landscape efficiently. In this research, we develop three emulations of a Chinese brush: a skeleton-based brush, a 2D texture footprint and a dynamic 3D footprint, all driven by the motion and pressure of a stylus pen. We propose a hybrid mapping to generate both the body and surface of volumetric clouds from the brush footprints. Our interface integrates these components along with 3D canvas control and GPU-based volumetric rendering into an interactive cloud modelling system. Our cloud modelling system is able to create various types of clouds occurring in nature. User tests indicate that our brush calligraphy approach is preferred to conventional volumetric cloud modelling and that it produces convincing 3D cloud formations in an intuitive and interactive fashion. While traditional modelling systems focus on surface generation of 3D objects, our brush calligraphy technique constructs the interior structure. This forms the basis of a new modelling style for objects with amorphous shape
Augmented reality (AR) for surgical robotic and autonomous systems: State of the art, challenges, and solutions
Despite the substantial progress achieved in the development and integration of augmented reality (AR) in surgical robotic and autonomous systems (RAS), the center of focus in most devices remains on improving end-effector dexterity and precision, as well as improved access to minimally invasive surgeries. This paper aims to provide a systematic review of different types of state-of-the-art surgical robotic platforms while identifying areas for technological improvement. We associate specific control features, such as haptic feedback, sensory stimuli, and human-robot collaboration, with AR technology to perform complex surgical interventions for increased user perception of the augmented world. Current researchers in the field have, for long, faced innumerable issues with low accuracy in tool placement around complex trajectories, pose estimation, and difficulty in depth perception during two-dimensional medical imaging. A number of robots described in this review, such as Novarad and SpineAssist, are analyzed in terms of their hardware features, computer vision systems (such as deep learning algorithms), and the clinical relevance of the literature. We attempt to outline the shortcomings in current optimization algorithms for surgical robots (such as YOLO and LTSM) whilst providing mitigating solutions to internal tool-to-organ collision detection and image reconstruction. The accuracy of results in robot end-effector collisions and reduced occlusion remain promising within the scope of our research, validating the propositions made for the surgical clearance of ever-expanding AR technology in the future
Shared control for natural motion and safety in hands-on robotic surgery
Hands-on robotic surgery is where the surgeon controls the tool's motion by applying forces and torques to the robot holding the tool, allowing the robot-environment interaction to be felt though the tool itself. To further improve results, shared control strategies are used to combine the strengths of the surgeon with those of the robot. One such strategy is active constraints, which prevent motion into regions deemed unsafe or unnecessary. While research in active constraints on rigid anatomy has been well-established, limited work on dynamic active constraints (DACs) for deformable soft tissue has been performed, particularly on strategies which handle multiple sensing modalities. In addition, attaching the tool to the robot imposes the end effector dynamics onto the surgeon, reducing dexterity and increasing fatigue. Current control policies on these systems only compensate for gravity, ignoring other dynamic effects. This thesis presents several research contributions to shared control in hands-on robotic surgery, which create a more natural motion for the surgeon and expand the usage of DACs to point clouds. A novel null-space based optimization technique has been developed which minimizes the end effector friction, mass, and inertia of redundant robots, creating a more natural motion, one which is closer to the feeling of the tool unattached to the robot. By operating in the null-space, the surgeon is left in full control of the procedure. A novel DACs approach has also been developed, which operates on point clouds. This allows its application to various sensing technologies, such as 3D cameras or CT scans and, therefore, various surgeries. Experimental validation in point-to-point motion trials and a virtual reality ultrasound scenario demonstrate a reduction in work when maneuvering the tool and improvements in accuracy and speed when performing virtual ultrasound scans. Overall, the results suggest that these techniques could increase the ease of use for the surgeon and improve patient safety.Open Acces
Markov-Gibbs Random Field Approach for Modeling of Skin Surface Textures
Medical imaging has been contributing to dermatology by providing computer-based
assistance by 2D digital imaging of skin and processing of images. Skin imaging can be more
effective by inclusion of 3D skin features. Furthermore, clinical examination of skin consists
of both visual and tactile inspection. The tactile sensation is related to 3D surface profiles and
mechanical parameters. The 3D imaging of skin can also be integrated with haptic
technology for computer-based tactile inspection. The research objective of this work is to
model 3D surface textures of skin. A 3D image acquisition set up capturing skin surface
textures at high resolution (~0.1 mm) has been used. An algorithm to extract 2D grayscale
texture (height map) from 3D texture has been presented. The extracted 2D textures are then
modeled using Markov-Gibbs random field (MGRF) modeling technique. The modeling
results for MGRF model depend on input texture characteristics. The homogeneous, spatially
invariant texture patterns are modeled successfully. From the observation of skin samples, we
classify three key features of3D skin profiles i.e. curvature of underlying limb, wrinkles/line
like features and fine textures. The skin samples are distributed in three input sets to see the
MGRF model's response to each of these 3D features. First set consists of all three features.
Second set is obtained after elimination of curvature and contains both wrinkle/line like
features and fine textures. Third set is also obtained after elimination of curvature but
consists of fine textures only.
MGRF modeling for set I did not result in any visual similarity. Hence the curvature of
underlying limbs cannot be modeled successfully and makes an inhomogeneous feature. For
set 2 the wrinkle/line like features can be modeled with low/medium visual similarity
depending on the spatial invariance. The results for set 3 show that fine textures of skin are
almost always modeled successfully with medium/high visual similarity and make a
homogeneous feature. We conclude that the MGRF model is able to model fine textures of
skin successfully which are on scale of~ 0.1 mm. The surface profiles on this resolution can
provide haptic sensation of roughness and friction. Therefore fine textures can be an
important clue to different skin conditions perceived through tactile inspection via a haptic
device
Evaluation of haptic guidance virtual fixtures and 3D visualization methods in telemanipulation—a user study
© 2019, The Author(s). This work presents a user-study evaluation of various visual and haptic feedback modes on a real telemanipulation platform. Of particular interest is the potential for haptic guidance virtual fixtures and 3D-mapping techniques to enhance efficiency and awareness in a simple teleoperated valve turn task. An RGB-Depth camera is used to gather real-time color and geometric data of the remote scene, and the operator is presented with either a monocular color video stream, a 3D-mapping voxel representation of the remote scene, or the ability to place a haptic guidance virtual fixture to help complete the telemanipulation task. The efficacy of the feedback modes is then explored experimentally through a user study, and the different modes are compared on the basis of objective and subjective metrics. Despite the simplistic task and numerous evaluation metrics, results show that the haptic virtual fixture resulted in significantly better collision avoidance compared to 3D visualization alone. Anticipated performance enhancements were also observed moving from 2D to 3D visualization. Remaining comparisons lead to exploratory inferences that inform future direction for focused and statistically significant studies
Simulating molecular docking with haptics
Intermolecular binding underlies various metabolic and regulatory processes of the
cell, and the therapeutic and pharmacological properties of drugs. Molecular docking
systems model and simulate these interactions in silico and allow the study of the
binding process. In molecular docking, haptics enables the user to sense the interaction
forces and intervene cognitively in the docking process. Haptics-assisted docking
systems provide an immersive virtual docking environment where the user can interact
with the molecules, feel the interaction forces using their sense of touch, identify
visually the binding site, and guide the molecules to their binding pose. Despite a
forty-year research e�ort however, the docking community has been slow to adopt this
technology. Proprietary, unreleased software, expensive haptic hardware and limits
on processing power are the main reasons for this. Another signi�cant factor is the
size of the molecules simulated, limited to small molecules.
The focus of the research described in this thesis is the development of an interactive
haptics-assisted docking application that addresses the above issues, and enables
the rigid docking of very large biomolecules and the study of the underlying interactions.
Novel methods for computing the interaction forces of binding on the CPU
and GPU, in real-time, have been developed. The force calculation methods proposed
here overcome several computational limitations of previous approaches, such as precomputed
force grids, and could potentially be used to model molecular
exibility
at haptic refresh rates. Methods for force scaling, multipoint collision response, and
haptic navigation are also reported that address newfound issues, particular to the
interactive docking of large systems, e.g. force stability at molecular collision. The
i
ii
result is a haptics-assisted docking application, Haptimol RD, that runs on relatively
inexpensive consumer level hardware, (i.e. there is no need for specialized/proprietary
hardware)
- …