54 research outputs found
Mobile-cloud assisted video summarization framework for efficient management of remote sensing data generated by wireless capsule sensors
YesWireless capsule endoscopy (WCE) has great advantages over traditional endoscopy
because it is portable and easy to use, especially in remote monitoring health-services.
However, during the WCE process, the large amount of captured video data demands a
significant deal of computation to analyze and retrieve informative video frames. In order to
facilitate efficient WCE data collection and browsing task, we present a resource- and
bandwidth-aware WCE video summarization framework that extracts the representative
keyframes of the WCE video contents by removing redundant and non-informative frames.
For redundancy elimination, we use Jeffrey-divergence between color histograms and
inter-frame Boolean series-based correlation of color channels. To remove non-informative
frames, multi-fractal texture features are extracted to assist the classification using an
ensemble-based classifier. Owing to the limited WCE resources, it is impossible for the
WCE system to perform computationally intensive video summarization tasks. To resolve
computational challenges, mobile-cloud architecture is incorporated, which provides resizable
computing capacities by adaptively offloading video summarization tasks between the client
and the cloud server. The qualitative and quantitative results are encouraging and show that
the proposed framework saves information transmission cost and bandwidth, as well as the
valuable time of data analysts in browsing remote sensing data.Supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2013R1A1A2012904)
Deep neural networks in the cloud: Review, applications, challenges and research directions
Deep neural networks (DNNs) are currently being deployed as machine learning technology in a wide
range of important real-world applications. DNNs consist of a huge number of parameters that require
millions of floating-point operations (FLOPs) to be executed both in learning and prediction modes. A
more effective method is to implement DNNs in a cloud computing system equipped with centralized
servers and data storage sub-systems with high-speed and high-performance computing capabilities.
This paper presents an up-to-date survey on current state-of-the-art deployed DNNs for cloud computing.
Various DNN complexities associated with different architectures are presented and discussed alongside
the necessities of using cloud computing. We also present an extensive overview of different cloud
computing platforms for the deployment of DNNs and discuss them in detail. Moreover, DNN applications
already deployed in cloud computing systems are reviewed to demonstrate the advantages of using
cloud computing for DNNs. The paper emphasizes the challenges of deploying DNNs in cloud computing
systems and provides guidance on enhancing current and new deployments.The EGIA project (KK-2022/00119The
Consolidated Research Group MATHMODE (IT1456-22
An m-health application for cerebral stroke detection and monitoring using cloud services
[EN] Over 25 million people suffered from cerebral strokes in a span of 23 years. Many systems are being developed to monitor and improve the life of patients that suffer from different diseases. However, solutions for cerebral strokes are hard to find. Moreover, due to their widespread utilization, smartphones have presented themselves as the most appropriate devices for many e-health systems. In this paper, we propose a cerebral stroke detection solution that employs the cloud to store and analyze data in order to provide statistics to public institutions. Moreover, the prototype of the application is presented. The three most important symptoms of cerebral strokes were considered to develop the tasks that are conducted. Thus, the first task detects smiles, the second task employs voice recognition to determine if a sentence is repeated correctly and, the third task determines if the arms can be raised. Several tests were performed in order to verify the application. Results show its ability to determine whether users have the symptoms of cerebral stroke or not.This work has been partially supported by the pre-doctoral student grant "Ayudas para contratos predoctorales de Formacion del Profesorado Universitario FPU (Convocatoria 2014)" by the "Ministerio de Educacion, Cultura y Deporte", with reference: FPU14/02953.García-García, L.; Tomás Gironés, J.; Parra-Boronat, L.; Lloret, J. (2019). An m-health application for cerebral stroke detection and monitoring using cloud services. International Journal of Information Management. 45:319-327. https://doi.org/10.1016/j.ijinfomgt.2018.06.004S3193274
Recommended from our members
Machine learning based small bowel video capsule endoscopy analysis: Challenges and opportunities
YesVideo capsule endoscopy (VCE) is a revolutionary technology for the early diagnosis of gastric disorders. However, owing to the high redundancy and subtle manifestation of anomalies among thousands of frames, the manual construal of VCE videos requires considerable patience, focus, and time. The automatic analysis of these videos using computational methods is a challenge as the capsule is untamed in motion and captures frames inaptly. Several machine learning (ML) methods, including recent deep convolutional neural networks approaches, have been adopted after evaluating their potential of improving the VCE analysis. However, the clinical impact of these methods is yet to be investigated. This survey aimed to highlight the gaps between existing ML-based research methodologies and clinically significant rules recently established by gastroenterologists based on VCE. A framework for interpreting raw frames into contextually relevant frame-level findings and subsequently merging these findings with meta-data to obtain a disease-level diagnosis was formulated. Frame-level findings can be more intelligible for discriminative learning when organized in a taxonomical hierarchy. The proposed taxonomical hierarchy, which is formulated based on pathological and visual similarities, may yield better classification metrics by setting inference classes at a higher level than training classes. Mapping from the frame level to the disease level was structured in the form of a graph based on clinical relevance inspired by the recent international consensus developed by domain experts. Furthermore, existing methods for VCE summarization, classification, segmentation, detection, and localization were critically evaluated and compared based on aspects deemed significant by clinicians. Numerous studies pertain to single anomaly detection instead of a pragmatic approach in a clinical setting. The challenges and opportunities associated with VCE analysis were delineated. A focus on maximizing the discriminative power of features corresponding to various subtle lesions and anomalies may help cope with the diverse and mimicking nature of different VCE frames. Large multicenter datasets must be created to cope with data sparsity, bias, and class imbalance. Explainability, reliability, traceability, and transparency are important for an ML-based diagnostics system in a VCE. Existing ethical and legal bindings narrow the scope of possibilities where ML can potentially be leveraged in healthcare. Despite these limitations, ML based video capsule endoscopy will revolutionize clinical practice, aiding clinicians in rapid and accurate diagnosis
Recommended from our members
Divide-and-conquer based summarization framework for extracting affective video content
YesRecent advances in multimedia technology have led to tremendous increases in the available volume of video data, thereby creating a major requirement for efficient systems to manage such huge data volumes. Video summarization is one of the key techniques for accessing and managing large video libraries. Video summarization can be used to extract the affective contents of a video sequence to generate a concise representation of its content. Human attention models are an efficient means of affective content extraction. Existing visual attention driven summarization frameworks have high computational cost and memory requirements, as well as a lack of efficiency in accurately perceiving human attention. To cope with these issues, we propose a divide-and-conquer based framework for an efficient summarization of big video data. We divide the original video data into shots, where an attention model is computed from each shot in parallel. Viewer's attention is based on multiple sensory perceptions, i.e., aural and visual, as well as the viewer's neuronal signals. The aural attention model is based on the Teager energy, instant amplitude, and instant frequency, whereas the visual attention model employs multi-scale contrast and motion intensity. Moreover, the neuronal attention is computed using the beta-band frequencies of neuronal signals. Next, an aggregated attention curve is generated using an intra- and inter-modality fusion mechanism. Finally, the affective content in each video shot is extracted. The fusion of multimedia and neuronal signals provides a bridge that links the digital representation of multimedia with the viewer’s perceptions. Our experimental results indicate that the proposed shot-detection based divide-and-conquer strategy mitigates the time and computational complexity. Moreover, the proposed attention model provides an accurate reflection of the user preferences and facilitates the extraction of highly affective and personalized summaries.Supported by the ICT R&D program of MSIP/IITP. [2014(R0112-14-1014), The Development of Open Platform for Service of Convergence Contents]
Efficient Deep CNN-Based Fire Detection and Localisation in Video Surveillance Applications
Convolutional neural networks (CNN) have yielded state-of-the-art performance in image classification and other computer vision tasks. Their application in fire detection systems will substantially improve detection accuracy, which will eventually minimize fire disasters and reduce the ecological and social ramifications. However, the major concern with CNN-based fire detection systems is their implementation in real-world surveillance networks, due to their high memory and computational requirements for inference. In this work, we propose an energy-friendly and computationally efficient CNN architecture, inspired by the SqueezeNet architecture for fire detection, localization, and semantic understanding of the scene of the fire. It uses smaller convolutional kernels and contains no dense, fully connected layers, which helps keep the computational requirements to a minimum. Despite its low computational needs, the experimental results demonstrate that our proposed solution achieves accuracies that are comparable to other, more complex models, mainly due to its increased depth. Moreover, the paper shows how a trade-off can be reached between fire detection accuracy and efficiency, by considering the specific characteristics of the problem of interest and the variety of fire data
- …