1,390 research outputs found

    An Epipolar Line from a Single Pixel

    Full text link
    Computing the epipolar geometry from feature points between cameras with very different viewpoints is often error prone, as an object's appearance can vary greatly between images. For such cases, it has been shown that using motion extracted from video can achieve much better results than using a static image. This paper extends these earlier works based on the scene dynamics. In this paper we propose a new method to compute the epipolar geometry from a video stream, by exploiting the following observation: For a pixel p in Image A, all pixels corresponding to p in Image B are on the same epipolar line. Equivalently, the image of the line going through camera A's center and p is an epipolar line in B. Therefore, when cameras A and B are synchronized, the momentary images of two objects projecting to the same pixel, p, in camera A at times t1 and t2, lie on an epipolar line in camera B. Based on this observation we achieve fast and precise computation of epipolar lines. Calibrating cameras based on our method of finding epipolar lines is much faster and more robust than previous methods.Comment: WACV 201

    Object Tracking: Appearance Modeling And Feature Learning

    Get PDF
    Object tracking in real scenes is an important problem in computer vision due to increasing usage of tracking systems day in and day out in various applications such as surveillance, security, monitoring and robotic vision. Object tracking is the process of locating objects of interest in every frame of video frames. Many systems have been proposed to address the tracking problem where the major challenges come from handling appearance variation during tracking caused by changing scale, pose, rotation, illumination and occlusion. In this dissertation, we address these challenges by introducing several novel tracking techniques. First, we developed a multiple object tracking system that deals specially with occlusion issues. The system depends on our improved KLT tracker for accurate and robust tracking during partial occlusion. In full occlusion, we applied a Kalman filter to predict the object\u27s new location and connect the trajectory parts. Many tracking methods depend on a rectangle or an ellipse mask to segment and track objects. Typically, using a larger or smaller mask will lead to loss of tracked objects. Second, we present an object tracking system (SegTrack) that deals with partial and full occlusions by employing improved segmentation methods: mixture of Gaussians and a silhouette segmentation algorithm. For re-identification, one or more feature vectors for each tracked object are used after target reappearing. Third, we propose a novel Bayesian Hierarchical Appearance Model (BHAM) for robust object tracking. Our idea is to model the appearance of a target as combination of multiple appearance models, each covering the target appearance changes under a certain situation (e.g. view angle). In addition, we built an object tracking system by integrating BHAM with background subtraction and the KLT tracker for static camera videos. For moving camera videos, we applied BHAM to cluster negative and positive target instances. As tracking accuracy depends mainly on finding good discriminative features to estimate the target location, finally, we propose to learn good features for generic object tracking using online convolutional neural networks (OCNN). In order to learn discriminative and stable features for tracking, we propose a novel object function to train OCNN by penalizing the feature variations in consecutive frames, and the tracker is built by integrating OCNN with a color-based multi-appearance model. Our experimental results on real-world videos show that our tracking systems have superior performance when compared with several state-of-the-art trackers. In the feature, we plan to apply the Bayesian Hierarchical Appearance Model (BHAM) for multiple objects tracking

    Pictures in Your Mind: Using Interactive Gesture-Controlled Reliefs to Explore Art

    Get PDF
    Tactile reliefs offer many benefits over the more classic raised line drawings or tactile diagrams, as depth, 3D shape, and surface textures are directly perceivable. Although often created for blind and visually impaired (BVI) people, a wider range of people may benefit from such multimodal material. However, some reliefs are still difficult to understand without proper guidance or accompanying verbal descriptions, hindering autonomous exploration. In this work, we present a gesture-controlled interactive audio guide (IAG) based on recent low-cost depth cameras that can be operated directly with the hands on relief surfaces during tactile exploration. The interactively explorable, location-dependent verbal and captioned descriptions promise rapid tactile accessibility to 2.5D spatial information in a home or education setting, to online resources, or as a kiosk installation at public places. We present a working prototype, discuss design decisions, and present the results of two evaluation studies: the first with 13 BVI test users and the second follow-up study with 14 test users across a wide range of people with differences and difficulties associated with perception, memory, cognition, and communication. The participant-led research method of this latter study prompted new, significant and innovative developments

    ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฐ˜ ๋‹จ์ผ ๊ฑฐ๋ฆฌ ๊ณต๊ฐ„ ๋‚ด GPCR ๋‹จ๋ฐฑ์งˆ๊ตฐ ๊ณ„์ธต ๊ตฌ์กฐ์˜ ๋™์‹œ์  ๋ชจ๋ธ๋ง ๊ธฐ๋ฒ•

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€,2019. 8. ๊น€์„ .G ๋‹จ๋ฐธ์งˆ ์—ฐ๊ฒฐ ์ˆ˜์šฉ์ฒด(GPCR)์€ ๊ณ„์ธต ๊ตฌ์กฐ๋กœ ํ˜•์„ฑ๋œ ๋‹ค์–‘ํ•œ ๋‹จ๋ฐฑ์งˆ๊ตฐ์œผ๋กœ ๊ตฌ์„ฑ๋œ๋‹ค. ๋‹จ๋ฐฑ์งˆ ์„œ์—ด์„ ํ†ตํ•œ GPCR์— ๋Œ€ํ•œ ๊ณ„์‚ฐ์ ์ธ ๋ชจ๋ธ๋ง์€ ๊ตฐ(family), ์•„๊ตฐ(subfamily), ์ค€์•„๊ตฐ(sub-subfamily)์˜ ๊ฐ ๊ณ„์ธต์—์„œ ๋…๋ฆฝ์ ์œผ๋กœ ์‹คํ–‰๋˜๋Š” ๋ฐฉ์‹์œผ๋กœ ์ด๋ฃจ์–ด์ ธ์™”๋‹ค. ํ•˜์ง€๋งŒ ์ด๋Ÿฌํ•œ ์ ‘๊ทผ ๋ฐฉ์‹๋“ค์€ ๋‹จ์ ˆ๋œ ๋ชจ๋ธ๋“ค์„ ํ†ตํ•˜์—ฌ ๋‹จ๋ฐฑ์งˆ ๋‚ด์˜ ์ •๋ณด๋ฅผ ์ฒ˜๋ฆฌํ•˜๊ธฐ ๋•Œ๋ฌธ์— GPCR ์ข…๋ฅ˜ ์‚ฌ์ด์˜ ๊ด€๊ณ„๋Š” ๊ณ ๋ คํ•˜์ง€ ๋ชปํ•œ๋‹ค๋Š” ํ•œ๊ณ„๋ฅผ ๊ฐ€์ง€๊ณ  ์žˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ์—์„œ๋Š” ๋”ฅ๋Ÿฌ๋‹์„ ์ด์šฉํ•˜์—ฌ GPCR์˜ ๊ณ„์ธต ๊ตฌ์กฐ์—์„œ ๋‚˜ํƒ€๋‚˜๋Š” ํŠน์ง•๋“ค์„ ๋‹จ์ผํ•œ ๋ชจ๋ธ๋กœ ๋™์‹œ์ ์œผ๋กœ ํ•™์Šตํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ์ œ์‹œํ•œ๋‹ค. ๋˜ํ•œ ๊ณ„์ธต์ ์ธ ๊ด€๊ณ„๋“ค์„ ํ•˜๋‚˜์˜ ๋ฒกํ„ฐ ๊ณต๊ฐ„์— ๊ฑฐ๋ฆฌ๋ฅผ ํ†ตํ•ด ํ‘œํ˜„ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๊ธฐ ์œ„ํ•œ ์†์‹คํ•จ์ˆ˜๋„ ์ œ์‹œํ•œ๋‹ค. ์ด ์—ฐ๊ตฌ๋Š” GPCR ์ˆ˜์šฉ์ฒด๋“ค์˜ ์—ฌ๋Ÿฌ ๊ณ„์ธต์—์„œ ๊ณตํ†ต์ ์œผ๋กœ ๋‚˜ํƒ€๋‚˜๋Š” ํŠน์ง•๋“ค์„ ํ•™์Šตํ•˜๊ณ  ํ‘œํ˜„ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ๋‹ค๋ฃจ๊ณ  ์žˆ๋‹ค. ์—ฌ๋Ÿฌ ์‹ฌํ™”์ ์ธ ์‹คํ—˜๋“ค์„ ํ†ตํ•˜์—ฌ ์šฐ๋ฆฌ๋Š” ๊ธฐ์ˆ ์ ์ธ ์ธก๋ฉด๊ณผ ์ƒ๋ฌผํ•™์ ์ธ ์ธก๋ฉด์—์„œ ๋‹จ๋ฐฑ์งˆ ๊ฐ„ ๊ณ„์ธต์ ์ธ ๊ด€๊ณ„๊ฐ€ ์„ฑ๊ณต์ ์œผ๋กœ ํ•™์Šต์ด ๋˜์—ˆ๋‹ค๋Š” ๊ฒƒ์„ ๋ณด์˜€๋‹ค. ์ฒซ๋ฒˆ์งธ๋กœ, ์šฐ๋ฆฌ๋Š” ์ž„๋ฒ ๋”ฉ ๋ฒกํ„ฐ์— ๊ณ„์ธต์  ๊ตฐ์ง‘ํ™”(hierarchical clustering) ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ ์šฉํ•จ์œผ๋กœ์จ ๊ณ„ํ†ต์ˆ˜(phylogenetic tree)๋ฅผ ๋งŒ๋“ค์—ˆ๊ณ , ๊ตฐ์ง‘ ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ์‹ค์ œ ๊ณ„์ธต ๊ตฌ์กฐ์™€์˜ ์ˆ˜์น˜์ ์ธ ๋น„๊ต๋ฅผ ํ†ตํ•˜์—ฌ ์ž„๋ฒ ๋”ฉ ๋ฒกํ„ฐ๋ฅผ ํ†ตํ•ด ๊ณ„ํ†ตํ•™์  ํŠน์ง•์— ๋Œ€ํ•œ ์œ ์ถ”๊ฐ€ ๊ฐ€๋Šฅํ•˜๋‹ค๋Š” ๊ฒƒ์„ ๋ณด์˜€๋‹ค. ๋‘๋ฒˆ์งธ๋กœ, ์ž„๋ฒ ๋”ฉ ๋ฒกํ„ฐ์˜ ๊ตฐ์ง‘ํ™” ๊ฒฐ๊ณผ์— ๋‹ค์ค‘ ์„œ์—ด ์ •๋ ฌ(multiple sequence alignment)๋ฅผ ์ ์šฉ์‹œํ‚ด์œผ๋กœ์จ ์ƒ๋ฌผํ•™์ ์œผ๋กœ ์œ ์˜๋ฏธํ•œ ์„œ์—ด์  ํŠน์„ฑ๋“ค์„ ์ฐพ์•„๋‚ผ ์ˆ˜ ์žˆ๋‹ค๋Š” ๊ฒƒ์„ ๋ณด์˜€๋‹ค. ์ด๋Š” ์ž„๋ฒ ๋”ฉ ๋ฒกํ„ฐ ๋ถ„์„์ด GPCR ๋‹จ๋ฐฑ์งˆ ์—ฐ๊ตฌ์— ์žˆ์–ด ํšจ์œจ์ ์ธ ์ฒซ๊ฑธ์Œ์ด ๋  ์ˆ˜ ์žˆ๋‹ค๋Š” ๊ฒƒ์„ ๋ณด์—ฌ์ค€๋‹ค. ์ด๋Ÿฌํ•œ ๊ฒฐ๊ณผ๋Š” ์—ฌ๋Ÿฌ ๊ณ„์ธต์œผ๋กœ ์ด๋ฃจ์–ด์ง„ ๋‹จ๋ฐฑ์งˆ๊ตฐ์— ๋Œ€ํ•œ ๋™์‹œ์ ์ธ ๋ชจ๋ธ๋ง์ด ๊ฐ€๋Šฅํ•˜๋‹ค๋Š” ๊ฒƒ์„ ๋งํ•˜๊ณ  ์žˆ๋‹ค.G protein-coupled receptors (GPCRs) belong to diverse families of proteins that can be defined at multiple levels. Computational modeling of GPCR families from the sequences has been performed separately at each level of family, sub-family, and sub-subfamily. However, relationships between classes are ignored in these approaches as they process the information in the sequences with a group of disconnected models. In this work, we propose a deep learning network to simultaneously learn representations in the GPCR hierarchy with a unified model and a loss term to express hierarchical relations in terms of distances in a single embedding space. The model introduces a method to learn and construct shared representations across hierarchies of the protein family. In extensive experiments, we showed that hierarchical relations between sequences are successfully captured in our model in both of technical and biological aspect. First, we showed that phylogenetic information in the sequences can be inferred from the vectors by constructing phylogenetic tree using hierarchical clustering algorithm and by quantitatively analyzing the quality of clustering results compared to the real label information. Second, inspection on embedding vectors is demonstrated to be a effective first step to-ward an analysis of GPCR proteins by showing that biologically significant sequence features can be revealed from multiple sequence alignments on clustering results on embedding vectors. Our work showed that simultaneous modeling of protein families with multiple hierarchies is possible.Abstract i Chapter โ… . Introduction 1 1.1 Background 1 1.2 Motivation 3 Chapter โ…ก. Methods 7 2.1 Data Preparation 7 2.1.1 Dataset 7 2.1.2 Data representation 7 2.2 Model architecture 8 2.2.1 Feature extractor with CNN 8 2.2.2 Embedding layer 8 2.2.3 Output layer 9 2.3 Loss function 10 2.3.1 Softmax loss 10 2.3.2 Center loss 10 2.3.3 Overall loss 12 2.4 Training procedure 13 2.5 Evaluation metric 14 2.5.1 Silhouette score 14 2.5.2 Adjusted mutual information score 15 Chapter โ…ข. Results 17 3.1 Evaluation on hierarchical structure 17 3.1.1 Preservation of distances 17 3.1.2 Phylogenetic tree reconstruction 20 3.1.3 Quantitative evaluation on clustering results 21 3.2 Sequence analysis with embedding vectors 26 3.2.1 Technical analysis 26 3.2.2 Biological analysis 28 3.3 Classification accuracy 30 Chapter โ…ฃ. Conclusion 32 References 35Maste

    Multi-view Performance Capture of Surface Details

    No full text

    Embedded System Object Tracking Using Webcam

    Get PDF
    The extensive availability of hardware devices and intensive expansion their computing power have been the catalyst behind the rapid development of computer vision. In this project, an implementation of object tracking in an inexpensive and small embedded system platform is presented. The tracking system comprised of two Raspberry Pis with two different cameras used: a webcam and Raspicam module. Three communication connection models of the system are discussed in this paper for establishing communication between the two Raspberry Pis. Data sharing between these two hardware platforms is the proposed solution for resolving the limited processing power each platform possesses. The SimpleCV, an open source framework that provides free computer vision libraries that is useful for object detection and tracking algorithm development

    Which One is Me?: Identifying Oneself on Public Displays

    Get PDF
    While user representations are extensively used on public displays, it remains unclear how well users can recognize their own representation among those of surrounding users. We study the most widely used representations: abstract objects, skeletons, silhouettes and mirrors. In a prestudy (N=12), we identify five strategies that users follow to recognize themselves on public displays. In a second study (N=19), we quantify the users' recognition time and accuracy with respect to each representation type. Our findings suggest that there is a significant effect of (1) the representation type, (2) the strategies performed by users, and (3) the combination of both on recognition time and accuracy. We discuss the suitability of each representation for different settings and provide specific recommendations as to how user representations should be applied in multi-user scenarios. These recommendations guide practitioners and researchers in selecting the representation that optimizes the most for the deployment's requirements, and for the user strategies that are feasible in that environment

    Vision-based traffic surveys in urban environments

    Get PDF
    This paper presents a state-of-the-art, vision-based vehicle detection and type classification to perform traffic surveys from a roadside closed-circuit television camera. Vehicles are detected using background subtraction based on a Gaussian mixture model that can cope with vehicles that become stationary over a significant period of time. Vehicle silhouettes are described using a combination of shape and appearance features using an intensity-based pyramid histogram of orientation gradients (HOG). Classification is performed using a support vector machine, which is trained on a small set of hand-labeled silhouette exemplars. These exemplars are identified using a model-based preclassifier that utilizes calibrated images mapped by Google Earth to provide accurately surveyed scene geometry matched to visible image landmarks. Kalman filters track the vehicles to enable classification by majority voting over several consecutive frames. The system counts vehicles and separates them into four categories: car, van, bus, and motorcycle (including bicycles). Experiments with real-world data have been undertaken to evaluate system performance and vehicle detection rates of 96.45% and classification accuracy of 95.70% have been achieved on this data.The authors gratefully acknowledge the Royal Borough of Kingston for providing the video data. S.A. Velastin is grateful to funding received from the Universidad Carlos III de Madrid, the European Unionโ€™s Seventh Framework Programme for research, technological development and demonstration under grant agreement nยบ 600371, el Ministerio de Economรญa y Competitividad (COFUND2013-51509) and Banco Santander
    • โ€ฆ
    corecore