28 research outputs found

    Non-linear Convolution Filters for CNN-based Learning

    Full text link
    During the last years, Convolutional Neural Networks (CNNs) have achieved state-of-the-art performance in image classification. Their architectures have largely drawn inspiration by models of the primate visual system. However, while recent research results of neuroscience prove the existence of non-linear operations in the response of complex visual cells, little effort has been devoted to extend the convolution technique to non-linear forms. Typical convolutional layers are linear systems, hence their expressiveness is limited. To overcome this, various non-linearities have been used as activation functions inside CNNs, while also many pooling strategies have been applied. We address the issue of developing a convolution method in the context of a computational model of the visual cortex, exploring quadratic forms through the Volterra kernels. Such forms, constituting a more rich function space, are used as approximations of the response profile of visual cells. Our proposed second-order convolution is tested on CIFAR-10 and CIFAR-100. We show that a network which combines linear and non-linear filters in its convolutional layers, can outperform networks that use standard linear filters with the same architecture, yielding results competitive with the state-of-the-art on these datasets.Comment: 9 pages, 5 figures, code link, ICCV 201

    Affect state recognition for adaptive human robot interaction in learning environments

    Get PDF
    Previous studies of robots used in learning environments suggest that the interaction between learner and robot is able to enhance the learning procedure towards a better engagement of the learner. Moreover, intelligent robots can also adapt their behavior during a learning process according to certain criteria resulting in increasing cognitive learning gains. Motivated by these results, we propose a novel Human Robot Interaction framework where the robot adjusts its behavior to the affect state of the learner. Our framework uses the theory of flow to label different affect states (i.e., engagement, boredom and frustration) and adapt the robot's actions. Based on the automatic recognition of these states, through visual cues, our method adapt the learning actions taking place at this moment and performed by the robot. This results in keeping the learner at most times engaged in the learning process. In order to recognizing the affect state of the user a two step approach is followed. Initially we recognize the facial expressions of the learner and therefore we map these to an affect state. Our algorithm perform well even in situations where the environment is noisy due to the presence of more than one person and/or situations where the face is partially occluded

    Comprehensive Comparison of Deep Learning Models for Lung and COVID-19 Lesion Segmentation in CT scans

    Full text link
    Recently there has been an explosion in the use of Deep Learning (DL) methods for medical image segmentation. However the field's reliability is hindered by the lack of a common base of reference for accuracy/performance evaluation and the fact that previous research uses different datasets for evaluation. In this paper, an extensive comparison of DL models for lung and COVID-19 lesion segmentation in Computerized Tomography (CT) scans is presented, which can also be used as a benchmark for testing medical image segmentation models. Four DL architectures (Unet, Linknet, FPN, PSPNet) are combined with 25 randomly initialized and pretrained encoders (variations of VGG, DenseNet, ResNet, ResNext, DPN, MobileNet, Xception, Inception-v4, EfficientNet), to construct 200 tested models. Three experimental setups are conducted for lung segmentation, lesion segmentation and lesion segmentation using the original lung masks. A public COVID-19 dataset with 100 CT scan images (80 for train, 20 for validation) is used for training/validation and a different public dataset consisting of 829 images from 9 CT scan volumes for testing. Multiple findings are provided including the best architecture-encoder models for each experiment as well as mean Dice results for each experiment, architecture and encoder independently. Finally, the upper bounds improvements when using lung masks as a preprocessing step or when using pretrained models are quantified. The source code and 600 pretrained models for the three experiments are provided, suitable for fine-tuning in experimental setups without GPU capabilities.Comment: 10 pages, 8 figures, 2 table

    NEW 3D METHOD TO ESTIMATE THE CYCLING FRONTAL AREA DURING PEDALLING

    Get PDF
    Aerodynamic drag represents a high percentage of the resistive forces in cycling. The frontal area has been traditionally measured through 2D digitalization, which requires a capture and an analysis processes. In the present study, the 2D method has been compared with a new real time 3D method to measure the frontal area in three different cycling positions. High reliability (ICC\u3e0.9) was found on both measurement methods. However there were differences between methods at the most aerodynamic positions. These differences might appear due to the capacity of the new method to measure the depth dimension, which supposes an improvement on the frontal area analysis. Furthermore, the new 3D method allows feedback of the frontal area in real time

    Using Tobit Kalman filtering in order to improve the Motion recorded by Microsoft Kinect

    Get PDF
    In this paper, we analyze data from Microsoft Kinect v2 camera using Kalman Tobit and Kalman filters so as to minimize noise. The data concern three-dimensional spatial coordinates recording movements of a persons’A joints, which are subject to measurement errors. The noise variances of the process and the measurements are estimated using the maximum likelihood function. In order to include into the model restrictive conditions based on anthropometric data (e.g. the distances between various joints) we apply the Tobit Kalman Filter. Additionally, restrictions for the joints displacements per fame based on real data can be used in order to get better results. Finally simulations of skeleton before and after using Kalman filtering are presented

    Two examples of online eHealth platforms for supporting people living with cognitive impairments and their caregivers

    Get PDF
    This paper compares two methodological approaches derived from the EU Horizon 2020 funded projects CAREGIVERSPRO-MMD (C-MMD)1 and ICT4LIFE2. Both approaches were initiated in 2016 with the ambition to provide new integrated care services to people living with cognitive impairments, including Dementia, Alzheimer and Parkinson disease, as well as to their home caregivers towards a long-term increase in quality of life and autonomy at home. An outline of the disparities and similarities related to non-pharmacological interventions introduced by the two projects to foster treatment adherence was made. Both approaches have developed software solutions, including social platforms, notifications, Serious Games, user monitoring and support services aimed at developing the concepts of self-care, active patients and integrated care. Besides their differences, both projects can be benefited by knowledge and technology exchange, pilot results sharing and possible user's exchange if possible in the near future.Peer ReviewedPostprint (published version

    Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks

    No full text
    The undeniable computational power of artificial neural networks has granted the scientific community the ability to exploit the available data in ways previously inconceivable. However, deep neural networks require an overwhelming quantity of data in order to interpret the underlying connections between them, and therefore, be able to complete the specific task that they have been assigned to. Feeding a deep neural network with vast amounts of data usually ensures efficiency, but may, however, harm the network’s ability to generalize. To tackle this, numerous regularization techniques have been proposed, with dropout being one of the most dominant. This paper proposes a selective gradient dropout method, which, instead of relying on dropping random weights, learns to freeze the training process of specific connections, thereby increasing the overall network’s sparsity in an adaptive manner, by driving it to utilize more salient weights. The experimental results show that the produced sparse network outperforms the baseline on numerous image classification datasets, and additionally, the yielded results occurred after significantly less training epochs

    Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks

    No full text
    The undeniable computational power of artificial neural networks has granted the scientific community the ability to exploit the available data in ways previously inconceivable. However, deep neural networks require an overwhelming quantity of data in order to interpret the underlying connections between them, and therefore, be able to complete the specific task that they have been assigned to. Feeding a deep neural network with vast amounts of data usually ensures efficiency, but may, however, harm the network’s ability to generalize. To tackle this, numerous regularization techniques have been proposed, with dropout being one of the most dominant. This paper proposes a selective gradient dropout method, which, instead of relying on dropping random weights, learns to freeze the training process of specific connections, thereby increasing the overall network’s sparsity in an adaptive manner, by driving it to utilize more salient weights. The experimental results show that the produced sparse network outperforms the baseline on numerous image classification datasets, and additionally, the yielded results occurred after significantly less training epochs
    corecore