2 research outputs found
Exploiting Social Networks. Technological Trends (Habilitation Thesis)
The habilitation thesis presents two main directions:
1. Exploiting data from social networks (Twitter, Facebook, Flickr, etc.) -
creating resources for text and image processing (classification, retrieval,
credibility, diversification, etc.);
2. Creating applications with new technologies : augmented reality
(eLearning, games, smart museums, gastronomy, etc.), virtual reality (eLearning
and games), speech processing with Amazon Alexa (eLearning, entertainment, IoT,
etc.).
The work was validated with good results in evaluation campaigns like CLEF
(Question Answering, Image CLEF, LifeCLEF, etc.), SemEval (Sentiment and
Emotion in text, Anorexia, etc.).Comment: 162 pages, This thesis presents the author's research activity after
March 2009 when he defended his Ph.D. Thesis "Textual Entailment" from the
artificial intelligence domain, related to natural language processing (NLP
UAIC participation at ImageCLEF 2012 Photo Annotation Task
Abstract. This paper presents the participation of our group in the ImageCLEF 2012 Photo Annotation Task. Our approach is based on visual and textual features as we experiment with different strategies in order to extract the semantics inside an image. First, we construct a textual dictionary of tags using the most frequent words present in the user tag annotated images from the training data sets. A linear kernel is then developed based on this dictionary. To gather more information from the images we further extract local and global visual features using TopSurf and Profile Entropy Features as well as Color Moments technique. We then aggregate these features with Support Vector Machines classification algorithm and train separate SVM models for each concept. In the end, to improve our system’s performance, we add a postprocessing step that verifies the consistency of the predicted concepts and also applies a face detection algorithm in order to increase the recognition accuracy of the person related concepts. Our submission consists of one visual-only and four multi-modal runs. We further give a more detailed perspective of our system and discuss our results and conclusions