15 research outputs found

    GTAutoAct: An Automatic Datasets Generation Framework Based on Game Engine Redevelopment for Action Recognition

    Full text link
    Current datasets for action recognition tasks face limitations stemming from traditional collection and generation methods, including the constrained range of action classes, absence of multi-viewpoint recordings, limited diversity, poor video quality, and labor-intensive manually collection. To address these challenges, we introduce GTAutoAct, a innovative dataset generation framework leveraging game engine technology to facilitate advancements in action recognition. GTAutoAct excels in automatically creating large-scale, well-annotated datasets with extensive action classes and superior video quality. Our framework's distinctive contributions encompass: (1) it innovatively transforms readily available coordinate-based 3D human motion into rotation-orientated representation with enhanced suitability in multiple viewpoints; (2) it employs dynamic segmentation and interpolation of rotation sequences to create smooth and realistic animations of action; (3) it offers extensively customizable animation scenes; (4) it implements an autonomous video capture and processing pipeline, featuring a randomly navigating camera, with auto-trimming and labeling functionalities. Experimental results underscore the framework's robustness and highlights its potential to significantly improve action recognition model training

    Markerless tumor-tracking algorithm using prior 4D-CBCT

    Get PDF
    Respiratory motion management is a huge challenge in radiation therapy. Respiratory motion induces temporal anatomic changes that distort the tumor volume and its position. In this study, a markerless tumor-tracking algorithm was investigated by performing phase recognition during stereotactic body radiation therapy (SBRT) using four-dimensional cone-beam computer tomography (4D-CBCT) obtained at patient registration, and in-treatment cone-beam projection images. The data for 20 treatment sessions (five lung cancer patients) were selected for this study. Three of the patients were treated with conventional flattening filter (FF) beams, and the other two were treated with flattening filter-free (FFF) beams. Prior to treatment, 4D-CBCT was acquired to create the template projection images for 10 phases. In-treatment images were obtained at near real time during treatment. Template-based phase recognition was performed for 4D-CBCT re-projected templates using prior 4D-CBCT based phase recognition algorithm and was compared with the results generated by the Amsterdam Shroud (AS) technique. Visual verification technique was used for the verification of the phase recognition and AS technique at certain tumor-visible angles. Offline template matching analysis using the cross-correlation indicated that phase recognition performed using the prior 4D-CBCT and visual verification matched up to 97.5% in the case of FFF, and 95% in the case of FF, whereas the AS technique matched 83.5% with visual verification for FFF and 93% for FF. Markerless tumor tracking based on phase recognition using prior 4D-CBCT has been developed successfully. This is the first study that reports on the use of prior 4D-CBCT based on normalized cross-correlation technique for phase recognition

    Analysis of Mesoscopic Electromagnetic Phenomena in TypeII Superconductors by the Fluxoid Dynamics Method

    No full text
    報告番号: 甲12606 ; 学位授与年月日: 1997-03-28 ; 学位の種別: 課程博士 ; 学位の種類: 博士(工学) ; 学位記番号: 博工第3884号 ; 研究科・専攻: 工学系研究科システム量子工学専

    A Vision-Based Approach for Ensuring Proper Use of Personal Protective Equipment (PPE) in Decommissioning of Fukushima Daiichi Nuclear Power Station

    No full text
    Decommissioning of the Fukushima Daiichi nuclear power station (NPS) is challenging due to industrial and chemical hazards as well as radiological ones. The decommissioning workers in these sites are instructed to wear proper Personal Protective Equipment (PPE) for radiation protection. However, workers may not be able to accurately comply with safety regulations at decommissioning sites, even with prior education and training. In response to the difficulties of on-site PPE management, this paper presents a vision-based automated monitoring approach to help to facilitate the occupational safety monitoring task of decommissioning workers to ensure proper use of PPE by the combination of deep learning-based individual detection and object detection using geometric relationships analysis. The performance of the proposed approach was experimentally evaluated, and the experimental results demonstrate that the proposed approach is capable of identifying decommissioning workers’ improper use of PPE with high precision and recall rate while ensuring real-time performance to meet the industrial requirements

    磁束量子動力学法による第二種超電導体中のメゾスコピック電磁現象の解析

    Get PDF
    University of Tokyo (東京大学

    Prediction of the motion of chest internal points using a recurrent neural network trained with real-time recurrent learning for latency compensation in lung cancer radiotherapy

    No full text
    During the radiotherapy treatment of patients with lung cancer, the radiation delivered to healthy tissue around the tumor needs to be minimized, which is difficult because of respiratory motion and the latency of linear accelerator (LINAC) systems. In the proposed study, we first use the Lucas-Kanade pyramidal optical flow algorithm to perform deformable image registration (DIR) of chest computed tomography (CT) scan images of four patients with lung cancer. We then track three internal points close to the lung tumor based on the previously computed deformation field and predict their position with a recurrent neural network (RNN) trained using real-time recurrent learning (RTRL) and gradient clipping. The breathing data is quite regular, sampled at approximately 2.5 Hz, and includes artificially added drift in the spine direction. The amplitude of the motion of the tracked points ranged from 12.0 mm to 22.7 mm. Finally, we propose a simple method for recovering and predicting three-dimensional (3D) tumor images from the tracked points and the initial tumor image, based on a linear correspondence model and the Nadaraya-Watson non-linear regression. The root-mean-square (RMS) error, maximum error and jitter corresponding to the RNN prediction on the test set were smaller than the same performance measures obtained with linear prediction and least mean squares (LMS). In particular, the maximum prediction error associated with the RNN, equal to 1.51 mm, is respectively 16.1% and 5.0% lower than the error given by a linear predictor and LMS. The average prediction time per time step with RTRL is equal to 119 ms, which is less than the 400 ms marker position sampling time. The tumor position in the predicted images appears visually correct, which is confirmed by the high mean cross-correlation between the original and predicted images, equal to 0.955. The standard deviation of the Gaussian kernel and the number of layers in the optical flow algorithm were the parameters having the most significant impact on registration performance. Their optimization led respectively to a 31.3% and 36.2% decrease in the registration error. Using only a single layer proved to be detrimental to the registration quality because tissue motion in the lower part of the lung has a high amplitude relative to the resolution of the CT scan images. The random initialization of the hidden units and the number of these hidden units were found to be the most important factors affecting the performance of the RNN. Increasing the number of hidden units from 15 to 250 led to a 56.3% decrease in the prediction error on the cross-validation data. Similarly, optimizing the standard deviation of the initial Gaussian distribution of the synaptic weights led to a 28.4% decrease in the prediction error on the cross-validation data, with the error minimized for with the four patients
    corecore