7,946 research outputs found

    A survey on Human Mobility and its applications

    Full text link
    Human Mobility has attracted attentions from different fields of studies such as epidemic modeling, traffic engineering, traffic prediction and urban planning. In this survey we review major characteristics of human mobility studies including from trajectory-based studies to studies using graph and network theory. In trajectory-based studies statistical measures such as jump length distribution and radius of gyration are analyzed in order to investigate how people move in their daily life, and if it is possible to model this individual movements and make prediction based on them. Using graph in mobility studies, helps to investigate the dynamic behavior of the system, such as diffusion and flow in the network and makes it easier to estimate how much one part of the network influences another by using metrics like centrality measures. We aim to study population flow in transportation networks using mobility data to derive models and patterns, and to develop new applications in predicting phenomena such as congestion. Human Mobility studies with the new generation of mobility data provided by cellular phone networks, arise new challenges such as data storing, data representation, data analysis and computation complexity. A comparative review of different data types used in current tools and applications of Human Mobility studies leads us to new approaches for dealing with mentioned challenges

    Growing a Tree in the Forest: Constructing Folksonomies by Integrating Structured Metadata

    Full text link
    Many social Web sites allow users to annotate the content with descriptive metadata, such as tags, and more recently to organize content hierarchically. These types of structured metadata provide valuable evidence for learning how a community organizes knowledge. For instance, we can aggregate many personal hierarchies into a common taxonomy, also known as a folksonomy, that will aid users in visualizing and browsing social content, and also to help them in organizing their own content. However, learning from social metadata presents several challenges, since it is sparse, shallow, ambiguous, noisy, and inconsistent. We describe an approach to folksonomy learning based on relational clustering, which exploits structured metadata contained in personal hierarchies. Our approach clusters similar hierarchies using their structure and tag statistics, then incrementally weaves them into a deeper, bushier tree. We study folksonomy learning using social metadata extracted from the photo-sharing site Flickr, and demonstrate that the proposed approach addresses the challenges. Moreover, comparing to previous work, the approach produces larger, more accurate folksonomies, and in addition, scales better.Comment: 10 pages, To appear in the Proceedings of ACM SIGKDD Conference on Knowledge Discovery and Data Mining(KDD) 201

    Fine-grained characterization of edge workloads

    Get PDF
    Edge computing is an emerging paradigm within the field of distributed computing, aimed at bringing data processing capabilities closer to the data-generating sources to enable real-time processing and reduce latency. However, the lack of representative data in the literature poses a significant challenge for evaluating the effectiveness of new algorithms and techniques developed for this paradigm. A part of the process towards alleviating this problem includes creating realistic and relevant workloads for the edge computing community. Research has already been conducted towards this goal, but resulting workload characterizations from these studies have been shown to not give an accurate representation of the workloads. This research gap highlights the need for developing new methodologies that can accurately characterize edge computing workloads. In this work we propose a novel methodology to characterize edge computing workloads, which leverages hardware performance counters to capture the behavior and characteristics of edge workloads in high detail. We explore the concept of representing workloads in a high-dimensional space, and develop a "proof-of-concept" classification model, that classifies workloads on a continuous "imprecise" data spectrum, to demonstrate the effectiveness and potential of the proposed characterization methodology. This research contributes to the field of edge computing by identifying and addressing the limitations of existing edge workload characterization techniques, and also opens up further avenues of research with regards to edge computing workload characterization

    Text categorization and similarity analysis: similarity measure, literature review

    Get PDF
    Document classification and provenance has become an important area of computer science as the amount of digital information is growing significantly. Organisations are storing documents on computers rather than in paper form. Software is now required that will show the similarities between documents (i.e. document classification) and to point out duplicates and possibly the history of each document (i.e. provenance). Poor organisation is common and leads to situations like above. There exists a number of software solutions in this area designed to make document organisation as simple as possible. I'm doing my project with Pingar who are a company based in Auckland who aim to help organise the growing amount of unstructured digital data. This reports analyses the existing literature in this area with the aim to determine what already exists and how my project will be different from existing solutions
    corecore