9 research outputs found
A computer vision approach to monitoring the activity and well-being of honeybees
Honeybees, in their role as pollinators, are vital to both agriculture and the wider ecosystem. However, they have experienced a serious decline across much of the world over recent years. Monitoring their well-being, and taking appropriate action if that is in jeopardy, has thus become a matter of great importance. In this paper, we present an approach based on computer vision to monitor bee activity and motion in the vicinity of an entrance/exit to a hive, including identifying and counting the number of bees approaching or leaving the hive in a given image frame or sequence of image frames
Asynchronous Bioplausible Neuron for Spiking Neural Networks for Event-Based Vision
Spiking Neural Networks (SNNs) offer a biologically inspired approach to
computer vision that can lead to more efficient processing of visual data with
reduced energy consumption. However, maintaining homeostasis within these
networks is challenging, as it requires continuous adjustment of neural
responses to preserve equilibrium and optimal processing efficiency amidst
diverse and often unpredictable input signals. In response to these challenges,
we propose the Asynchronous Bioplausible Neuron (ABN), a dynamic spike firing
mechanism to auto-adjust the variations in the input signal. Comprehensive
evaluation across various datasets demonstrates ABN's enhanced performance in
image classification and segmentation, maintenance of neural equilibrium, and
energy efficiency.Comment: 10 page
Asynchronous Events-based Panoptic Segmentation using Graph Mixer Neural Network
In the context of robotic grasping, object segmentation encounters several
difficulties when faced with dynamic conditions such as real-time operation,
occlusion, low lighting, motion blur, and object size variability. In response
to these challenges, we propose the Graph Mixer Neural Network that includes a
novel collaborative contextual mixing layer, applied to 3D event graphs formed
on asynchronous events. The proposed layer is designed to spread spatiotemporal
correlation within an event graph at four nearest neighbor levels parallelly.
We evaluate the effectiveness of our proposed method on the Event-based
Segmentation (ESD) Dataset, which includes five unique image degradation
challenges, including occlusion, blur, brightness, trajectory, scale variance,
and segmentation of known and unknown objects. The results show that our
proposed approach outperforms state-of-the-art methods in terms of mean
intersection over the union and pixel accuracy. Code available at:
https://github.com/sanket0707/GNN-Mixer.gitComment: 9 pages, 6 figure
Bimodal SegNet : fused instance segmentation using events and RGB frames
Object segmentation enhances robotic grasping by aiding object identification. Complex environments and dynamic conditions pose challenges such as occlusion, low light conditions, motion blur and object size variance. To address these challenges, we propose a Bimodal SegNet that fuses two types of visual signals, event-based data and RGB frame data. The proposed Bimodal SegNet network has two distinct encoders — one for RGB signal input and another for Event signal input, in addition to an Atrous Pyramidal Feature Amplification module. Encoders capture and fuse the rich contextual information from different resolutions via a Cross-Domain Contextual Attention layer while the decoder obtains sharp object boundaries. The evaluation of the proposed method undertakes five unique image degradation challenges including occlusion, blur, brightness, trajectory and scale variance on the Event-based Segmentation (ESD) Dataset. The results show a 4%–6% MIOU score improvement over state-of-the-art methods in terms of mean intersection over the union and pixel accuracy. The source code, dataset and model are publicly available at: https://github.com/sanket0707/Bimodal-SegNet
A neuromorphic dataset for tabletop object segmentation in indoor cluttered environment
Event-based cameras are commonly leveraged to mitigate issues such as motion blur, low dynamic range, and limited time sampling, which plague conventional cameras. However, a lack of dedicated event-based datasets for benchmarking segmentation algorithms, especially those offering critical depth information for occluded scenes, has been observed. In response, this paper introduces a novel Event-based Segmentation Dataset (ESD), a high-quality event 3D spatial-temporal dataset designed for indoor object segmentation within cluttered environments. ESD encompasses 145 sequences featuring 14,166 manually annotated RGB frames, along with a substantial event count of 21.88 million and 20.80 million events from two stereo-configured event-based cameras. Notably, this densely annotated 3D spatial-temporal event-based segmentation benchmark for tabletop objects represents a pioneering initiative, providing event-wise depth, and annotated instance labels, in addition to corresponding RGBD frames. By releasing ESD, our aim is to offer the research community a challenging segmentation benchmark of exceptional quality
ESD-1
Event-based Segmentation Dataset (ESD), a high-quality 3D spatial and temporal dataset for object segmentation in an indoor cluttered environment.
MATLAB source code repository:
yellow07200/ESD_labeling_tool (github.com
ESD-2
Event-based Segmentation Dataset (ESD), a high-quality 3D spatial and temporal dataset for object segmentation in an indoor cluttered environment.
MATLAB source code repository:
yellow07200/ESD_labeling_tool (github.com