1,124 research outputs found

    Technology of swallowable capsule for medical applications

    Get PDF
    Medical technology has undergone major breakthroughs in recent years, especially in the area of the examination tools for diagnostic purposes. This paper reviews the swallowable capsule technology in the examination of the gastrointestinal system for various diseases. The wireless camera pill has created a more advanced method than many traditional examination methods for the diagnosis of gastrointestinal diseases such as gastroscopy by the use of an endoscope. After years of great innovation, commercial swallowable pills have been produced and applied in clinical practice. These smart pills can cover the examination of the gastrointestinal system and not only provide to the physicians a lot more useful data that is not available from the traditional methods, but also eliminates the use of the painful endoscopy procedure. In this paper, the key state-of-the-art technologies in the existing Wireless Capsule Endoscopy (WCE) systems are fully reported and the recent research progresses related to these technologies are reviewed. The paper ends by further discussion on the current technical bottlenecks and future research in this area

    Design and Development of an Automatic Blood Detection System for Capsule Endoscopy Images

    Full text link
    Wireless Capsule Endoscopy is a technique that allows for observation of the entire gastrointestinal tract in an easy and non-invasive way. However, its greatest limitation lies in the time required to analyze the large number of images generated in each examination for diagnosis, which is about 2 hours. This causes not only a high cost, but also a high probability of a wrong diagnosis due to the physician’s fatigue, while the variable appearance of abnormalities requires continuous concentration. In this work, we designed and developed a system capable of automatically detecting blood based on classification of extracted regions, following two different classification approaches. The first method consisted in extraction of hand-crafted features that were used to train machine learning algorithms, specifically Support Vector Machines and Random Forest, to create models for classifying images as healthy tissue or blood. The second method consisted in applying deep learning techniques, concretely convolutional neural networks, capable of extracting the relevant features of the image by themselves. The best results (95.7% sensitivity and 92.3% specificity) were obtained for a Random Forest model trained with features extracted from the histograms of the three HSV color space channels. For both methods we extracted square patches of several sizes using a sliding window, while for the first approach we also implemented the waterpixels technique in order to improve the classification resultsThis work was funded by the European Unions H2020: MSCA: ITN program for the “Wireless In-body Environment Communication WiBEC” project under the grant agreement no. 675353. Additionally, we gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan V GPU used for this research.Pons Suñer, P.; Noorda, R.; Nevárez, A.; Colomer, A.; Pons Beltrán, V.; Naranjo, V. (2019). Design and Development of an Automatic Blood Detection System for Capsule Endoscopy Images. En Lecture Notes in Artificial Intelligence. Springer. 105-113. https://doi.org/10.1007/978-3-030-33617-2_12S105113Berens, J., Finlayson, G.D., Qiu, G.: Image indexing using compressed colour histograms. IEE Proc. Vis., Image Signal Process. 147(4), 349–355 (2000). https://doi.org/10.1049/ip-vis:20000630Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001). https://doi.org/10.1023/A:1010933404324Buscaglia, J.M., et al.: Performance characteristics of the suspected blood indicator feature in capsule endoscopy according to indication for study. Clin. Gastroenterol. Hepatol. 6(3), 298–301 (2008). https://doi.org/10.1016/j.cgh.2007.12.029Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995). https://doi.org/10.1007/BF00994018Li, B., Meng, M.Q.H.: Computer-aided detection of bleeding regions for capsule endoscopy images. IEEE Trans. Biomed. Eng. 56(4), 1032–1039 (2009). https://doi.org/10.1109/TBME.2008.2010526Machairas, V., Faessel, M., Cárdenas-Peña, D., Chabardes, T., Walter, T., Decencière, E.: Waterpixels. IEEE Trans. Image Process. 24(11), 3707–3716 (2015). https://doi.org/10.1109/TIP.2015.2451011Novozámskỳ, A., Flusser, J., Tachecí, I., Sulík, L., Bureš, J., Krejcar, O.: Automatic blood detection in capsule endoscopy video. J. Biomed. Opt. 21(12), 126007 (2016). https://doi.org/10.1117/1.JBO.21.12.126007Signorelli, C., Villa, F., Rondonotti, E., Abbiati, C., Beccari, G., de Franchis, R.: Sensitivity and specificity of the suspected blood identification system in video capsule enteroscopy. Endoscopy 37(12), 1170–1173 (2005). https://doi.org/10.1055/s-2005-870410Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)Varma, S., Simon, R.: Bias in error estimation when using cross-validation for model selection. BMC Bioinform. 7(1), 91 (2006). https://doi.org/10.1186/1471-2105-7-9

    High concordance between trained nurses and gastroenterologists in evaluating recordings of small bowel video capsule endoscopy (VCE)

    Get PDF
    Background & Aims: The video capsule endoscopy (VCE) is an accurate and validated tool to investigate the entire small bowel mucosa, but VCE recordings interpretation by the gastroenterologist is time-consuming. A pre-reading of VCE recordings by an expert nurse could be accurate and cost saving. We assessed the concordance between nurses and gastroenterologists in detecting lesions on VCE examinations. Methods: This was a prospective study enrolling consecutive patients who had undergone VCE in clinical practice. Two trained nurses and two expert gastroenterologists participated in the study. At VCE pre-reading the nurses selected any abnormalities, saved them as “thumbnails” and classified the detected lesions as a vascular abnormality, ulcerative lesion, polyp, tumor mass, and unclassified lesion. Then, the gastroenterologist evaluated and interpreted the selected lesions and, successively, reviewed the entire video for potential missed lesions. The time for VCE evaluation was recorded. Results: A total of 95 VCE procedures performed on consecutive patients (M/F: 47/48; mean age: 63 ± 12 years, range: 27−86 years) were evaluated. Overall, the nurses detected at least one lesion in 54 (56.8%) patients. There was total agreement between nurses and gastroenterologists, no missing lesions being discovered at a second look of the entire VCE recording by the physician. The pre-reading procedure by nurse allowed a time reduction of medical evaluation from 49 (33-69) to 10 (8-16) minutes (difference:-79.6%). Conclusions: Our data suggest that trained nurses can accurately identify and select relevant lesions in thumbnails that subsequently were faster reviewed by the gastroenterologist for a final diagnosis. This could significantly reduce the cost of VCE procedure

    Detection of Intestinal Bleeding in Wireless Capsule Endoscopy using Machine Learning Techniques

    Get PDF
    Gastrointestinal (GI) bleeding is very common in humans, which may lead to fatal consequences. GI bleeding can usually be identified using a flexible wired endoscope. In 2001, a newer diagnostic tool, wireless capsule endoscopy (WCE) was introduced. It is a swallow-able capsule-shaped device with a camera that captures thousands of color images and wirelessly sends those back to a data recorder. After that, the physicians analyze those images in order to identify any GI abnormalities. But it takes a longer screening time which may increase the danger of the patients in emergency cases. It is therefore necessary to use a real-time detection tool to identify bleeding in the GI tract. Each material has its own spectral ‘signature’ which shows distinct characteristics in specific wavelength of light [33]. Therefore, by evaluating the optical characteristics, the presence of blood can be detected. In the study, three main hardware designs were presented: one using a two-wavelength based optical sensor and others using two six-wavelength based spectral sensors with AS7262 and AS7263 chips respectively to determine the optical characteristics of the blood and non-blood samples. The goal of the research is to develop a machine learning model to differentiate blood samples (BS) and non-blood samples (NBS) by exploring their optical properties. In this experiment, 10 levels of crystallized bovine hemoglobin solutions were used as BS and 5 food colors (red, yellow, orange, tan and pink) with different concentrations totaling 25 non-blood samples were used as NBS. These blood and non-blood samples were also combined with pig’s intestine to mimic in-vivo experimental environment. The collected samples were completely separated into training and testing data. Different spectral features are analyzed to obtain the optical information about the samples. Based on the performance on the selected most significant features of the spectral wavelengths, k-nearest neighbors algorithm (k-NN) is finally chosen for the automated bleeding detection. The proposed k-NN classifier model has been able to distinguish the BS and NBS with an accuracy of 91.54% using two wavelengths features and around 89% using three combined wavelengths features in the visible and near-infrared spectral regions. The research also indicates that it is possible to deploy tiny optical detectors to detect GI bleeding in a WCE system which could eliminate the need of time-consuming image post-processing steps

    Intelligent Hemorrhage Identification in Wireless Capsule Endoscopy Pictures Using AI Techniques.

    Get PDF
    Image segmentation in medical images is performed to extract valuable information from the images by concentrating on the region of interest. Mostly, the number of medical images generated from a diagnosis is large and not ideal to treat with traditional ways of segmentation using machine learning models due to their numerous and complex features. To obtain crucial features from this large set of images, deep learning is a good choice over traditional machine learning algorithms. Wireless capsule endoscopy images comprise normal and sick frames and often suffers with a big data imbalance ratio which is sometimes 1000:1 for normal and sick classes. They are also special type of confounding images due to movement of the (capsule) camera, organs and variations in luminance to capture the site texture inside the body. So, we have proposed an automatic deep learning model based to detect bleeding frames out of the WCE images. The proposed model is based on Convolutional Neural Network (CNN) and its performance is compared with state-of- the-art methods including Logistic Regression, Support Vector Machine, Artificial Neural Network and Random Forest. The proposed model reduces the computational burden by offering the automatic feature extraction. It has promising accuracy with an F1 score of 0.76

    Magnetic-Visual Sensor Fusion-based Dense 3D Reconstruction and Localization for Endoscopic Capsule Robots

    Full text link
    Reliable and real-time 3D reconstruction and localization functionality is a crucial prerequisite for the navigation of actively controlled capsule endoscopic robots as an emerging, minimally invasive diagnostic and therapeutic technology for use in the gastrointestinal (GI) tract. In this study, we propose a fully dense, non-rigidly deformable, strictly real-time, intraoperative map fusion approach for actively controlled endoscopic capsule robot applications which combines magnetic and vision-based localization, with non-rigid deformations based frame-to-model map fusion. The performance of the proposed method is demonstrated using four different ex-vivo porcine stomach models. Across different trajectories of varying speed and complexity, and four different endoscopic cameras, the root mean square surface reconstruction errors 1.58 to 2.17 cm.Comment: submitted to IROS 201
    corecore