7,946 research outputs found
A survey on Human Mobility and its applications
Human Mobility has attracted attentions from different fields of studies such
as epidemic modeling, traffic engineering, traffic prediction and urban
planning. In this survey we review major characteristics of human mobility
studies including from trajectory-based studies to studies using graph and
network theory. In trajectory-based studies statistical measures such as jump
length distribution and radius of gyration are analyzed in order to investigate
how people move in their daily life, and if it is possible to model this
individual movements and make prediction based on them. Using graph in mobility
studies, helps to investigate the dynamic behavior of the system, such as
diffusion and flow in the network and makes it easier to estimate how much one
part of the network influences another by using metrics like centrality
measures. We aim to study population flow in transportation networks using
mobility data to derive models and patterns, and to develop new applications in
predicting phenomena such as congestion. Human Mobility studies with the new
generation of mobility data provided by cellular phone networks, arise new
challenges such as data storing, data representation, data analysis and
computation complexity. A comparative review of different data types used in
current tools and applications of Human Mobility studies leads us to new
approaches for dealing with mentioned challenges
Growing a Tree in the Forest: Constructing Folksonomies by Integrating Structured Metadata
Many social Web sites allow users to annotate the content with descriptive
metadata, such as tags, and more recently to organize content hierarchically.
These types of structured metadata provide valuable evidence for learning how a
community organizes knowledge. For instance, we can aggregate many personal
hierarchies into a common taxonomy, also known as a folksonomy, that will aid
users in visualizing and browsing social content, and also to help them in
organizing their own content. However, learning from social metadata presents
several challenges, since it is sparse, shallow, ambiguous, noisy, and
inconsistent. We describe an approach to folksonomy learning based on
relational clustering, which exploits structured metadata contained in personal
hierarchies. Our approach clusters similar hierarchies using their structure
and tag statistics, then incrementally weaves them into a deeper, bushier tree.
We study folksonomy learning using social metadata extracted from the
photo-sharing site Flickr, and demonstrate that the proposed approach addresses
the challenges. Moreover, comparing to previous work, the approach produces
larger, more accurate folksonomies, and in addition, scales better.Comment: 10 pages, To appear in the Proceedings of ACM SIGKDD Conference on
Knowledge Discovery and Data Mining(KDD) 201
Fine-grained characterization of edge workloads
Edge computing is an emerging paradigm within the field of distributed computing, aimed at bringing data processing capabilities closer to the data-generating sources to enable real-time processing and reduce latency. However, the lack of representative data in the literature poses a significant challenge for evaluating the effectiveness of new algorithms and techniques developed for this paradigm.
A part of the process towards alleviating this problem includes creating realistic and relevant workloads for the edge computing community. Research has already been conducted towards this goal, but resulting workload characterizations from these studies have been shown to not give an accurate representation of the workloads. This research gap highlights the need for developing new methodologies that can accurately characterize edge computing workloads.
In this work we propose a novel methodology to characterize edge computing workloads, which leverages hardware performance counters to capture the behavior and characteristics of edge workloads in high detail. We explore the concept of representing workloads in a high-dimensional space, and develop a "proof-of-concept" classification model, that classifies workloads on a continuous "imprecise" data spectrum, to demonstrate the effectiveness and potential of the proposed characterization methodology.
This research contributes to the field of edge computing by identifying and addressing the limitations of existing edge workload characterization techniques, and also opens up further avenues of research with regards to edge computing workload characterization
Text categorization and similarity analysis: similarity measure, literature review
Document classification and provenance has become an important area of computer science as the amount of digital information is growing significantly. Organisations are storing documents on computers rather than in paper form. Software is now required that will show the similarities between documents (i.e. document classification) and to point out duplicates and possibly the history of each document (i.e. provenance). Poor organisation is common and leads to situations like above. There exists a number of software solutions in this area designed to make document organisation as simple as possible. I'm doing my project with Pingar who are a company based in Auckland who aim to help organise the growing amount of unstructured digital data. This reports analyses the existing literature in this area with the aim to determine what already exists and how my project will be different from existing solutions
- …