26,391 research outputs found
Relevance of ASR for the Automatic Generation of Keywords Suggestions for TV programs
Semantic access to multimedia content in audiovisual archives is to a large extent dependent on quantity and quality of the metadata, and particularly the content descriptions that are attached to the individual items. However, given the growing amount of materials that are being created on a daily basis and the digitization of existing analogue collections, the traditional manual annotation of collections puts heavy demands on resources, especially for large audiovisual archives. One way to address this challenge, is to introduce (semi) automatic annotation techniques for generating and/or enhancing metadata. The NWO funded CATCH-CHOICE project has investigated the extraction of keywords form textual resources related to the TV programs to be archived (context documents), in collaboration with the Dutch audiovisual archives, Sound and Vision. Besides the descriptions of the programs published by the broadcasters on their Websites, Automatic Speech Transcription (ASR) techniques from the CATCH-CHoral project, also provide textual resources that might be relevant for suggesting keywords. This paper investigates the suitability of ASR for generating such keywords, which we evaluate against manual annotations of the documents and against keywords automatically generated from context documents
The global hydrology education resource
This article is a selective overview of a range of contemporary teaching resources currently available globally for university hydrology educators, with an emphasis on web-based resources. Major governmental and scientific organizations relevant to the promotion of hydrology teaching are briefly introduced. Selected online teaching materials are then overviewed, i.e. PowerPoint presentations, course materials, and multimedia. A range of websites offering free basic hydrology modelling software are mentioned, together with some data file sources which could be used for teaching. Websites offering a considerable range of general hydrology links are also noted, as are websites providing international and national data sets which might be incorporated into teaching exercises. Finally, some discussion is given on reference material for different modes of hydrology teaching, including laboratory and field exercises
Recommended from our members
MUSCLE movie-database: a multimodal corpus with rich annotation for dialogue and saliency detection
Automatic Recognition of Mammal Genera on Camera-Trap Images using Multi-Layer Robust Principal Component Analysis and Mixture Neural Networks
The segmentation and classification of animals from camera-trap images is due
to the conditions under which the images are taken, a difficult task. This work
presents a method for classifying and segmenting mammal genera from camera-trap
images. Our method uses Multi-Layer Robust Principal Component Analysis (RPCA)
for segmenting, Convolutional Neural Networks (CNNs) for extracting features,
Least Absolute Shrinkage and Selection Operator (LASSO) for selecting features,
and Artificial Neural Networks (ANNs) or Support Vector Machines (SVM) for
classifying mammal genera present in the Colombian forest. We evaluated our
method with the camera-trap images from the Alexander von Humboldt Biological
Resources Research Institute. We obtained an accuracy of 92.65% classifying 8
mammal genera and a False Positive (FP) class, using automatic-segmented
images. On the other hand, we reached 90.32% of accuracy classifying 10 mammal
genera, using ground-truth images only. Unlike almost all previous works, we
confront the animal segmentation and genera classification in the camera-trap
recognition. This method shows a new approach toward a fully-automatic
detection of animals from camera-trap images
AudioPairBank: Towards A Large-Scale Tag-Pair-Based Audio Content Analysis
Recently, sound recognition has been used to identify sounds, such as car and
river. However, sounds have nuances that may be better described by
adjective-noun pairs such as slow car, and verb-noun pairs such as flying
insects, which are under explored. Therefore, in this work we investigate the
relation between audio content and both adjective-noun pairs and verb-noun
pairs. Due to the lack of datasets with these kinds of annotations, we
collected and processed the AudioPairBank corpus consisting of a combined total
of 1,123 pairs and over 33,000 audio files. One contribution is the previously
unavailable documentation of the challenges and implications of collecting
audio recordings with these type of labels. A second contribution is to show
the degree of correlation between the audio content and the labels through
sound recognition experiments, which yielded results of 70% accuracy, hence
also providing a performance benchmark. The results and study in this paper
encourage further exploration of the nuances in audio and are meant to
complement similar research performed on images and text in multimedia
analysis.Comment: This paper is a revised version of "AudioSentibank: Large-scale
Semantic Ontology of Acoustic Concepts for Audio Content Analysis
- âŠ