81,931 research outputs found
A Framework to Enable the Semantic Inferencing and Querying of Multimedia Content
Cultural institutions, broadcasting companies, academic, scientific and defence organisations are producing vast quantities of digital multimedia content. With this growth in audiovisual material comes the need for standardised representations encapsulating the rich semantic meaning required to enable the automatic filtering, machine processing, interpretation and assimilation of multimedia resources. Additionally generating high-level descriptions is difficult and manual creation is expensive although significant progress has been made in recent years on automatic segmentation and low-level feature recognition for multimedia. Within this paper we describe the application of semantic web technologies to enable the generation of high-level, domain-specific, semantic descriptions of multimedia content from low-level, automatically-extracted features. By applying the knowledge reasoning capabilities provided by ontologies and inferencing rules to large, multimedia data sets generated by scientific research communities, we hope to expedite solutions to the complex scientific problems they face
Strategies for Searching Video Content with Text Queries or Video Examples
The large number of user-generated videos uploaded on to the Internet
everyday has led to many commercial video search engines, which mainly rely on
text metadata for search. However, metadata is often lacking for user-generated
videos, thus these videos are unsearchable by current search engines.
Therefore, content-based video retrieval (CBVR) tackles this metadata-scarcity
problem by directly analyzing the visual and audio streams of each video. CBVR
encompasses multiple research topics, including low-level feature design,
feature fusion, semantic detector training and video search/reranking. We
present novel strategies in these topics to enhance CBVR in both accuracy and
speed under different query inputs, including pure textual queries and query by
video examples. Our proposed strategies have been incorporated into our
submission for the TRECVID 2014 Multimedia Event Detection evaluation, where
our system outperformed other submissions in both text queries and video
example queries, thus demonstrating the effectiveness of our proposed
approaches
Movies Tags Extraction Using Deep Learning
Retrieving information from movies is becoming increasingly
demanding due to the enormous amount of multimedia
data generated each day. Not only it helps in efficient
search, archiving and classification of movies, but is also instrumental
in content censorship and recommendation systems.
Extracting key information from a movie and summarizing
it in a few tags which best describe the movie presents
a dedicated challenge and requires an intelligent approach
to automatically analyze the movie. In this paper, we formulate
movies tags extraction problem as a machine learning
classification problem and train a Convolution Neural Network
(CNN) on a carefully constructed tag vocabulary. Our
proposed technique first extracts key frames from a movie
and applies the trained classifier on the key frames. The
predictions from the classifier are assigned scores and are
filtered based on their relative strengths to generate a compact
set of most relevant key tags. We performed a rigorous
subjective evaluation of our proposed technique for a
wide variety of movies with different experiments. The evaluation
results presented in this paper demonstrate that our
proposed approach can efficiently extract the key tags of a
movie with a good accuracy
- …