32,848 research outputs found

    Determining the Best Sensing Coverage for 2-Dimensional Acoustic Target Tracking

    Get PDF
    Distributed acoustic target tracking is an important application area of wireless sensor networks. In this paper we use algebraic geometry to formally model 2-dimensional acoustic target tracking and then prove its best degree of required sensing coverage. We present the necessary conditions for three sensing coverage to accurately compute the spatio-temporal information of a target object. Simulations show that 3-coverage accurately locates a target object only in 53% of cases. Using 4-coverage, we present two different methods that yield correct answers in almost all cases and have time and memory usage complexity of Θ(1). Analytic 4-coverage tracking is our first proposed method that solves a simultaneous equation system using the sensing information of four sensor nodes. Redundant answer fusion is our second proposed method that solves at least two sets of simultaneous equations of target tracking using the sensing information of two different sets of three sensor nodes, and fusing the results using a new customized formal majority voter. We prove that 4-coverage guarantees accurate 2-dimensional acoustic target tracking under ideal conditions

    Array of sensors: A spatiotemporal-state-space model for target trajectory tracking

    Get PDF
    In this paper, with the objective of tracking the trajectory of multiple mobile targets, a novel spatiotemporal-state-space model is introduced for an array of sensors distributed in space. Under the wideband assumption, the proposed model incorporates the array geometry in conjunction with crucial target parameters namely (i) ranges, (ii) directions, (iii) velocities and (iv) associated Doppler effects. Computer simulation studies show some representative examples where the proposed model is utilised to track the locations of sources in space with a very high accuracy

    Towards retrieving force feedback in robotic-assisted surgery: a supervised neuro-recurrent-vision approach

    Get PDF
    Robotic-assisted minimally invasive surgeries have gained a lot of popularity over conventional procedures as they offer many benefits to both surgeons and patients. Nonetheless, they still suffer from some limitations that affect their outcome. One of them is the lack of force feedback which restricts the surgeon's sense of touch and might reduce precision during a procedure. To overcome this limitation, we propose a novel force estimation approach that combines a vision based solution with supervised learning to estimate the applied force and provide the surgeon with a suitable representation of it. The proposed solution starts with extracting the geometry of motion of the heart's surface by minimizing an energy functional to recover its 3D deformable structure. A deep network, based on a LSTM-RNN architecture, is then used to learn the relationship between the extracted visual-geometric information and the applied force, and to find accurate mapping between the two. Our proposed force estimation solution avoids the drawbacks usually associated with force sensing devices, such as biocompatibility and integration issues. We evaluate our approach on phantom and realistic tissues in which we report an average root-mean square error of 0.02 N.Peer ReviewedPostprint (author's final draft
    • 

    corecore