116 research outputs found

    高速ビジョンを用いたリアルタイムビデオモザイキングと安定化に関する研究

    Get PDF
    広島大学(Hiroshima University)博士(工学)Doctor of Engineeringdoctora

    AirConstellations: In-Air Device Formations for Cross-Device Interaction via Multiple Spatially-Aware Armatures

    Get PDF
    AirConstellations supports a unique semi-fixed style of cross-device interactions via multiple self-spatially-aware armatures to which users can easily attach (or detach) tablets and other devices. In particular, AirConstellations affords highly flexible and dynamic device formations where the users can bring multiple devices together in-air - with 2-5 armatures poseable in 7DoF within the same workspace - to suit the demands of their current task, social situation, app scenario, or mobility needs. This affords an interaction metaphor where relative orientation, proximity, attaching (or detaching) devices, and continuous movement into and out of ad-hoc ensembles can drive context-sensitive interactions. Yet all devices remain self-stable in useful configurations even when released in mid-air. We explore flexible physical arrangement, feedforward of transition options, and layering of devices in-air across a variety of multi-device app scenarios. These include video conferencing with flexible arrangement of the person-space of multiple remote participants around a shared task-space, layered and tiled device formations with overview+detail and shared-to-personal transitions, and flexible composition of UI panels and tool palettes across devices for productivity applications. A preliminary interview study highlights user reactions to AirConstellations, such as for minimally disruptive device formations, easier physical transitions, and balancing "seeing and being seen"in remote work

    A Human-Embodied Drone for Dexterous Aerial Manipulation

    Full text link
    Current drones perform a wide variety of tasks in surveillance, photography, agriculture, package delivery, etc. However, these tasks are performed passively without the use of human interaction. Aerial manipulation shifts this paradigm and implements drones with robotic arms that allow interaction with the environment rather than simply sensing it. For example, in construction, aerial manipulation in conjunction with human interaction could allow operators to perform several tasks, such as hosing decks, drill into surfaces, and sealing cracks via a drone. This integration with drones will henceforth be known as dexterous aerial manipulation. Our recent work integrated the worker’s experience into aerial manipulation using haptic technology. The net effect was such a system could enable the worker to leverage drones and complete tasks while utilizing haptics on the task site remotely. However, the tasks were completed within the operator’s line-of-sight. Until now, immersive AR/VR frameworks has rarely been integrated in aerial manipulation. Yet, such a framework allows the drones to embody and transport the operator’s senses, actions, and presence to a remote location in real-time. As a result, the operator can both physically interact with the environment and socially interact with actual workers on the worksite. This dissertation presents a human-embodied drone interface for dexterous aerial manipulation. Using VR/AR technology, the interface allows the operator to leverage their intelligence to collaboratively perform desired tasks anytime, anywhere with a drone that possesses great dexterity

    Towards High-Frequency Tracking and Fast Edge-Aware Optimization

    Full text link
    This dissertation advances the state of the art for AR/VR tracking systems by increasing the tracking frequency by orders of magnitude and proposes an efficient algorithm for the problem of edge-aware optimization. AR/VR is a natural way of interacting with computers, where the physical and digital worlds coexist. We are on the cusp of a radical change in how humans perform and interact with computing. Humans are sensitive to small misalignments between the real and the virtual world, and tracking at kilo-Hertz frequencies becomes essential. Current vision-based systems fall short, as their tracking frequency is implicitly limited by the frame-rate of the camera. This thesis presents a prototype system which can track at orders of magnitude higher than the state-of-the-art methods using multiple commodity cameras. The proposed system exploits characteristics of the camera traditionally considered as flaws, namely rolling shutter and radial distortion. The experimental evaluation shows the effectiveness of the method for various degrees of motion. Furthermore, edge-aware optimization is an indispensable tool in the computer vision arsenal for accurate filtering of depth-data and image-based rendering, which is increasingly being used for content creation and geometry processing for AR/VR. As applications increasingly demand higher resolution and speed, there exists a need to develop methods that scale accordingly. This dissertation proposes such an edge-aware optimization framework which is efficient, accurate, and algorithmically scales well, all of which are much desirable traits not found jointly in the state of the art. The experiments show the effectiveness of the framework in a multitude of computer vision tasks such as computational photography and stereo.Comment: PhD thesi

    Scene Mapping and Understanding by Robotic Vision

    Get PDF
    The first mechanical Automaton concept was found in a Chinese text written in the 3rd century BC, while Computer Vision was born in the late 1960s. Therefore, visual perception applied to machines (i.e. the Machine Vision) is a young and exciting alliance. When robots came in, the new field of Robotic Vision was born, and these terms began to be erroneously interchanged. In short, we can say that Machine Vision is an engineering domain, which concern the industrial use of Vision. The Robotic Vision, instead, is a research field that tries to incorporate robotics aspects in computer vision algorithms. Visual Servoing, for example, is one of the problems that cannot be solved by computer vision only. Accordingly, a large part of this work deals with boosting popular Computer Vision techniques by exploiting robotics: e.g. the use of kinematics to localize a vision sensor, mounted as the robot end-effector. The remainder of this work is dedicated to the counterparty, i.e. the use of computer vision to solve real robotic problems like grasping objects or navigate avoiding obstacles. Will be presented a brief survey about mapping data structures most widely used in robotics along with SkiMap, a novel sparse data structure created both for robotic mapping and as a general purpose 3D spatial index. Thus, several approaches to implement Object Detection and Manipulation, by exploiting the aforementioned mapping strategies, will be proposed, along with a completely new Machine Teaching facility in order to simply the training procedure of modern Deep Learning networks

    Information-theoretic environment modeling for mobile robot localization

    Full text link
    To enhance robotic computational efficiency without degenerating accuracy, it is imperative to fit the right and exact amount of information in its simplest form to the investigated task. This thesis conforms to this reasoning in environment model building and robot localization. It puts forth an approach towards building maps and localizing a mobile robot efficiently with respect to unknown, unstructured and moderately dynamic environments. For this, the environment is modeled on an information-theoretic basis, more specifically in terms of its transmission property. Subsequently, the presented environment model, which does not specifically adhere to classical geometric modeling, succeeds in solving the environment disambiguation effectively. The proposed solution lays out a two-level hierarchical structure for localization. The structure makes use of extracted features, which are stored in two different resolutions in a single hybrid feature-map. This enables dual coarse-topological and fine-geometric localization modalities. The first level in the hierarchy describes the environment topologically, where a defined set of places is described by a probabilistic feature representation. A conditional entropy-based criterion is proposed to quantify the transinformation between the feature and the place domains. This criterion provides a double benefit of pruning the large dimensional feature space, and at the same time selecting the best discriminative features that overcome environment aliasing problems. Features with the highest transinformation are filtered and compressed to form a coarse resolution feature-map (codebook). Localization at this level is conducted through place matching. In the second level of the hierarchy, the map is viewed in high-resolution, as consisting of non-compressed entropy-processed features. These features are additionally tagged with their position information. Given the identified topological place provided by the first level, fine localization corresponding to the second level is executed using feature triangulation. To enhance the triangulation accuracy, redundant features are used and two metric evaluating criteria are employ-ed; one for dynamic features and mismatches detection, and another for feature selection. The proposed approach and methods have been tested in realistic indoor environments using a vision sensor and the Scale Invariant Feature Transform local feature extraction. Through experiments, it is demonstrated that an information-theoretic modeling approach is highly efficient in attaining combined accuracy and computational efficiency performances for localization. It has also been proven that the approach is capable of modeling environments with a high degree of unstructuredness, perceptual aliasing, and dynamic variations (illumination conditions; scene dynamics). The merit of employing this modeling type is that environment features are evaluated quantitatively, while at the same time qualitative conclusions are generated about feature selection and performance in a robot localization task. In this way, the accuracy of localization can be adapted in accordance with the available resources. The experimental results also show that the hybrid topological-metric map provides sufficient information to localize a mobile robot on two scales, independent of the robot motion model. The codebook exhibits fast and accurate topological localization at significant compression ratios. The hierarchical localization framework demonstrates robustness and optimized space and time complexities. This, in turn, provides scalability to large environments application and real-time employment adequacies

    Realization of a demonstrator slave for robotic minimally invasive surgery

    Get PDF
    Robots for Minimally Invasive Surgery (MIS) can improve the surgeon’s work conditions with respect to conventional MIS and to enable MIS with more complex procedures. This requires to provide the surgeon with tactile feedback to feel forces executed on e.g. tissue and sutures, which is partially lost in conventional MIS. Additionally use of a robot should improve the approach possibilities of a target organ by means of instrument degrees of freedom (DoFs) and of the entry points with a compact set-up. These requirements add to the requirements set by the most common commercially available system, the da Vinci which are: (i) dexterity, (ii) natural hand-eye coordination, (iii) a comfortable body posture, (iv) intuitive utilization, and (v) a stereoscopic ’3D’ view of the operation site. The purpose of Sofie (Surgeon’s operating force-feedback interface Eindhoven) is to evaluate the possible benefit of force-feedback and the approach of both patient and target organ. Sofie integrates master, slave, electronic hardware and control. This thesis focusses on the design and realization of a technology demonstrator of the Slave. To provide good accuracy and valuable force-feedback, good dynamic behavior and limited hysteresis are required. To this end the Slave includes (i) a relatively short force-path between its instrument-tips and between tip and patient, and (ii) a passive instrument-support by means of a remote kinematically fixed point of rotation. The incision tissue does not support the instrument. The Slave is connected directly to the table. It provides a 20 DoF highly adaptable stiff frame (pre-surgical set-up) with a short force-path between the instrumenttips and between instrument-tip and patient. During surgery this frame supports three 4 DoF manipulators, two for exchangeable 4 DoF instruments and one for an endoscope. The pre-surgical set-up of the Slave consists of a 5 DoF platform-adjustment with a platform. This platform can hold three 5 DoF manipulator-adjustments in line-up. The set-up is compact to avoid interference with the team, entirely mechanical and allows fast manual adjustment and fixation of the joints. It provides a stiff frame during surgery. A weight-compensation mechanism for the platformadjustment has been proposed. Measurements indicate all natural frequencies are above 25 Hz. The manipulator moves the instrument in 4 DoFs (??, , ?? and Z). Each manipulator passively supports its instrument with a parallelogram mechanism, providing a kinematically fixed point of rotation. Two manipulators have been designed in consecutive order. The first manipulator drives with a worm-wormwheel, the second design uses a ball-screw drive. This ball-screw drive reduces friction, which is preferred for next generations of the manipulator, since the worm-wormwheel drive shows a relatively low coherence at low frequencies. The compact ??Zmanipulator moves the instrument in ?? by rotating a drum. Friction wheels in the drum provide Z. Eventually, the drum will be removable from the manipulator for sterilization. This layout of the manipulator results in a small motion-envelope and least obstructs the team at the table. Force sensors measuring forces executed with the instrument, are integrated in the manipulator instead of at the instrument tip, to avoid all risks of electrical signals being introduced into the patient. Measurements indicate the separate sensors function properly. Integrated in the manipulator the sensors provide a good indication of the force but do suffer from some hysteresis which might be caused by moving wires. The instrument as realized consists of a drive-box, an instrument-tube and a 4 DoF tip. It provides the surgeon with three DoFs additional to the gripper of conventional MIS instruments. These DoFs include two lateral rotations (pitch and pivot) to improve the approach possibilities and the roll DoF will contribute in stitching. Pitch and roll are driven by means of bevelgears, driven with concentric tubes. Cables drive the pivot and close DoFs of the gripper. The transmissions are backdriveable for safety. Theoretical torques that can be achieved with this instrument approximate the requirements closely. Further research needs to reveal the torques achieved in practice and whether the requirements obtained from literature actually are required for these 4 DoF instruments. Force-sensors are proposed and can be integrated. Sofie currently consists of a master prototype with two 5 DoF haptic interfaces, the Slave and an electronic hardware cabinet. The surgeon uses the haptic interfaces of the Master to manipulate the manipulators and instruments of the Slave, while the actuated DoFs of the Master provide the surgeon with force-feedback. This project resulted in a demonstrator of the slave with force sensors incorporated, compact for easy approach of the patient and additional DoFs to increase approach possibilities of the target organ. This slave and master provide a good starting point to implement haptic controllers. These additional features may ultimately benefit both surgeon and patient
    corecore