2,785 research outputs found
Hoodsquare: Modeling and Recommending Neighborhoods in Location-based Social Networks
Information garnered from activity on location-based social networks can be
harnessed to characterize urban spaces and organize them into neighborhoods. In
this work, we adopt a data-driven approach to the identification and modeling
of urban neighborhoods using location-based social networks. We represent
geographic points in the city using spatio-temporal information about
Foursquare user check-ins and semantic information about places, with the goal
of developing features to input into a novel neighborhood detection algorithm.
The algorithm first employs a similarity metric that assesses the homogeneity
of a geographic area, and then with a simple mechanism of geographic
navigation, it detects the boundaries of a city's neighborhoods. The models and
algorithms devised are subsequently integrated into a publicly available,
map-based tool named Hoodsquare that allows users to explore activities and
neighborhoods in cities around the world.
Finally, we evaluate Hoodsquare in the context of a recommendation
application where user profiles are matched to urban neighborhoods. By
comparing with a number of baselines, we demonstrate how Hoodsquare can be used
to accurately predict the home neighborhood of Twitter users. We also show that
we are able to suggest neighborhoods geographically constrained in size, a
desirable property in mobile recommendation scenarios for which geographical
precision is key.Comment: ASE/IEEE SocialCom 201
Fast, scalable, Bayesian spike identification for multi-electrode arrays
We present an algorithm to identify individual neural spikes observed on
high-density multi-electrode arrays (MEAs). Our method can distinguish large
numbers of distinct neural units, even when spikes overlap, and accounts for
intrinsic variability of spikes from each unit. As MEAs grow larger, it is
important to find spike-identification methods that are scalable, that is, the
computational cost of spike fitting should scale well with the number of units
observed. Our algorithm accomplishes this goal, and is fast, because it
exploits the spatial locality of each unit and the basic biophysics of
extracellular signal propagation. Human intervention is minimized and
streamlined via a graphical interface. We illustrate our method on data from a
mammalian retina preparation and document its performance on simulated data
consisting of spikes added to experimentally measured background noise. The
algorithm is highly accurate
Cluster analysis for physical oceanographic data and oceanographic surveys in Turkish seas
Cluster analysis is a useful data mining method to obtain detailed information on the physical state of the ocean. The primary objective of this study is the development of a new spatio-temporal density-based algorithm for clustering physical oceanographic data. This study extends the regular spatial cluster analysis to deal with spatial data at different epochs. It also presents the sensitivity of the new algorithm to different parameter settings. The purpose of the sensitivity analysis presented in this paper is to identify the response of the algorithm to variations in input parameter values and boundary conditions. In order to demonstrate the usage of the new algorithm, this paper presents two oceanographic applications that cluster the sea-surface temperature (SST) and the sea-surface height residual (SSH) data which records the satellite observations of the Turkish Seas. It also evaluates and justifies the clustering results by using a cluster validation technique
Machine Learning Based Real-Time Quantification of Production from Individual Clusters in Shale Wells
Over the last two decades, there has been advances in downhole monitoring in oil and gas wells with the use of Fiber-Optic sensing technology such as the Distributed Temperature Sensing (DTS). Unlike a conventional production log that provides only snapshots of the well performance, DTS provides continuous temperature measurements along the entire wellbore.
Whether by fluid extraction or injection, oil and gas production changes reservoir conditions, and continuous monitoring of downhole conditions is highly desirable. This research study presents a tool for real-time quantification of production from individual perforation clusters in a multi-stage shale well using Artificial Intelligence and Machine Learning. The technique presented provides continuous production log on demand thereby providing opportunities for the optimization of completions design and hydraulic fracture treatments of future planned wells. A Fiber-Optic sensing enabled horizontal well MIP-3H in the Marcellus Shale has been selected for this work. MIP-3H is a 28-stage horizontal well drilled in July 2015, as part of a Department of Energy (DOE)-sponsored project - Marcellus Shale Energy & Environment Laboratory (MSEEL). A one-day conventional production logging operation has been performed on MIP-3H using a flow scanner while the installed Fiber-Optic DTS unit has collected temperature measurements every three hours along the well since completion. An ensemble of machine learning models has been developed using as input the DTS measurements taken during the production logging operation, details of mechanical logs, completions design and hydraulic fracture treatments data of the well to develop the real-time shale gas production monitoring tool
Transient motion classification through turbid volumes via parallelized single-photon detection and deep contrastive embedding
Fast noninvasive probing of spatially varying decorrelating events, such as
cerebral blood flow beneath the human skull, is an essential task in various
scientific and clinical settings. One of the primary optical techniques used is
diffuse correlation spectroscopy (DCS), whose classical implementation uses a
single or few single-photon detectors, resulting in poor spatial localization
accuracy and relatively low temporal resolution. Here, we propose a technique
termed Classifying Rapid decorrelation Events via Parallelized single photon
dEtection (CREPE)}, a new form of DCS that can probe and classify different
decorrelating movements hidden underneath turbid volume with high sensitivity
using parallelized speckle detection from a pixel SPAD array. We
evaluate our setup by classifying different spatiotemporal-decorrelating
patterns hidden beneath a 5mm tissue-like phantom made with rapidly
decorrelating dynamic scattering media. Twelve multi-mode fibers are used to
collect scattered light from different positions on the surface of the tissue
phantom. To validate our setup, we generate perturbed decorrelation patterns by
both a digital micromirror device (DMD) modulated at multi-kilo-hertz rates, as
well as a vessel phantom containing flowing fluid. Along with a deep
contrastive learning algorithm that outperforms classic unsupervised learning
methods, we demonstrate our approach can accurately detect and classify
different transient decorrelation events (happening in 0.1-0.4s) underneath
turbid scattering media, without any data labeling. This has the potential to
be applied to noninvasively monitor deep tissue motion patterns, for example
identifying normal or abnormal cerebral blood flow events, at multi-Hertz rates
within a compact and static detection probe.Comment: Journal submissio
Density-based spatial clustering and ordering points approach for characterizations of tourist behaviour
Knowledge about the spots where tourist activity is undertaken, including which segments from the tourist market visit them, is valuable information for tourist service managers. Nowadays, crowdsourced smartphones applications are used as part of tourist surveys looking for knowledge about the tourist in all phases of their journey. However, the representativeness of this type of source, or how to validate the outcomes, are part of the issues that still need to be solved. In this research, a method to discover hotspots using clustering techniques and give to these hotspots a data-driven interpretation is proposed. The representativeness of the dataset and the validation of the results against existing statistics is assessed. The method was evaluated using 124,725 trips, which have been gathered by 1505 devices. The results show that the proposed approach successfully detects hotspots related with the most common activities developed by overnight tourists and repeat visitors in the region under study
Application of Spatiotemporal Fuzzy C-Means Clustering for Crime Spot Detection
The various sources generate large volume of spatiotemporal data of different types including crime events. In order to detect crime spot and predict future events, their analysis is important. Crime events are spatiotemporal in nature; therefore a distance function is defined for spatiotemporal events and is used in Fuzzy C-Means algorithm for crime analysis. This distance function takes care of both spatial and temporal components of spatiotemporal data. We adopt sum of squared error (SSE) approach and Dunn index to measure the quality of clusters. We also perform the experimentation on real world crime data to identify spatiotemporal crime clusters.
- …