158 research outputs found
UW-ProCCaps: UnderWater Progressive Colourisation with Capsules
Underwater images are fundamental for studying and understanding the status
of marine life. We focus on reducing the memory space required for image
storage while the memory space consumption in the collecting phase limits the
time lasting of this phase leading to the need for more image collection
campaigns. We present a novel machine-learning model that reconstructs the
colours of underwater images from their luminescence channel, thus saving 2/3
of the available storage space. Our model specialises in underwater colour
reconstruction and consists of an encoder-decoder architecture. The encoder is
composed of a convolutional encoder and a parallel specialised classifier
trained with webly-supervised data. The encoder and the decoder use layers of
capsules to capture the features of the entities in the image. The colour
reconstruction process recalls the progressive and the generative adversarial
training procedures. The progressive training gives the ground for a generative
adversarial routine focused on the refining of colours giving the image bright
and saturated colours which bring the image back to life. We validate the model
both qualitatively and quantitatively on four benchmark datasets. This is the
first attempt at colour reconstruction in greyscale underwater images.
Extensive results on four benchmark datasets demonstrate that our solution
outperforms state-of-the-art (SOTA) solutions. We also demonstrate that the
generated colourisation enhances the quality of images compared to enhancement
models at the SOTA
Wild animals' biologging through machine learning models
In recent decades the biodiversity crisis has been characterised by a decline and extinction of many animal species worldwide. To aid in understanding the threats and causes of this demise, conservation scientists rely on remote assessments. Innovation in technology in the form of microelectromechanical systems (MEMs) has brought about great leaps forward in understanding of animal life. The MEMs are now readily available to ecologists for remotely monitoring the activities of wild animals. Since the advent of electronic tags, methods such as biologging are being increasingly applied to the study of animal ecology, providing information unattainable through other techniques. In this paper, we discuss a few relevant instances of biologging studies. We present an overview on biologging research area, describing the evolution of acquisition of behavioural information and the improvement provided by tags. In second part we will review some common data analysis techniques used to identify daily activity of animals
Comparison between transformers and convolutional models for fine-grained classification of insects
Fine-grained classification is challenging due to the difficulty of finding
discriminatory features. This problem is exacerbated when applied to
identifying species within the same taxonomical class. This is because species
are often sharing morphological characteristics that make them difficult to
differentiate. We consider the taxonomical class of Insecta. The identification
of insects is essential in biodiversity monitoring as they are one of the
inhabitants at the base of many ecosystems. Citizen science is doing brilliant
work of collecting images of insects in the wild giving the possibility to
experts to create improved distribution maps in all countries. We have billions
of images that need to be automatically classified and deep neural network
algorithms are one of the main techniques explored for fine-grained tasks. At
the SOTA, the field of deep learning algorithms is extremely fruitful, so how
to identify the algorithm to use? We focus on Odonata and Coleoptera orders,
and we propose an initial comparative study to analyse the two best-known layer
structures for computer vision: transformer and convolutional layers. We
compare the performance of T2TViT, a fully transformer-base, EfficientNet, a
fully convolutional-base, and ViTAE, a hybrid. We analyse the performance of
the three models in identical conditions evaluating the performance per
species, per morph together with sex, the inference time, and the overall
performance with unbalanced datasets of images from smartphones. Although we
observe high performances with all three families of models, our analysis shows
that the hybrid model outperforms the fully convolutional-base and fully
transformer-base models on accuracy performance and the fully transformer-base
model outperforms the others on inference speed and, these prove the
transformer to be robust to the shortage of samples and to be faster at
inference time
Human activity recognition using multisensor data fusion based on Reservoir Computing
Activity recognition plays a key role in providing activity assistance and care for users in smart homes. In this work, we present an activity recognition system that classifies in the near real-time a set of common daily activities exploiting both the data sampled by sensors embedded in a smartphone carried out by the user and the reciprocal Received Signal Strength (RSS) values coming from worn wireless sensor devices and from sensors deployed in the environment. In order to achieve an effective and responsive classification, a decision tree based on multisensor data-stream is applied fusing data coming from embedded sensors on the smartphone and environmental sensors before processing the RSS stream. To this end, we model the RSS stream, obtained from a Wireless Sensor Network (WSN), using Recurrent Neural Networks (RNNs) implemented as efficient Echo State Networks (ESNs), within the Reservoir Computing (RC) paradigm. We targeted the system for the EvAAL scenario, an international competition that aims at establishing benchmarks and evaluation metrics for comparing Ambient Assisted Living (AAL) solutions. In this paper, the performance of the proposed activity recognition system is assessed on a purposely collected real-world dataset, taking also into account a competitive neural network approach for performance comparison. Our results show that, with an appropriate configuration of the information fusion chain, the proposed system reaches a very good accuracy with a low deployment cost
Localizing Tortoise Nests by Neural Networks
The goal of this research is to recognize the nest digging activity of tortoises using a device mounted atop the tortoise carapace. The device classifies tortoise movements in order to discriminate between nest digging, and non-digging activity (specifically walking and eating). Accelerometer data was collected from devices attached to the carapace of a number of tortoises during their two-month nesting period. Our system uses an accelerometer and an activity recognition system (ARS) which is modularly structured using an artificial neural network and an output filter. For the purpose of experiment and comparison, and with the aim of minimizing the computational cost, the artificial neural network has been modelled according to three different architectures based on the input delay neural network (IDNN). We show that the ARS can achieve very high accuracy on segments of data sequences, with an extremely small neural network that can be embedded in programmable low power devices. Given that digging is typically a long activity (up to two hours), the application of ARS on data segments can be repeated over time to set up a reliable and efficient system, called Tortoise@, for digging activity recognition
- …