6,374 research outputs found
Foundational principles for large scale inference: Illustrations through correlation mining
When can reliable inference be drawn in the "Big Data" context? This paper
presents a framework for answering this fundamental question in the context of
correlation mining, with implications for general large scale inference. In
large scale data applications like genomics, connectomics, and eco-informatics
the dataset is often variable-rich but sample-starved: a regime where the
number of acquired samples (statistical replicates) is far fewer than the
number of observed variables (genes, neurons, voxels, or chemical
constituents). Much of recent work has focused on understanding the
computational complexity of proposed methods for "Big Data." Sample complexity
however has received relatively less attention, especially in the setting when
the sample size is fixed, and the dimension grows without bound. To
address this gap, we develop a unified statistical framework that explicitly
quantifies the sample complexity of various inferential tasks. Sampling regimes
can be divided into several categories: 1) the classical asymptotic regime
where the variable dimension is fixed and the sample size goes to infinity; 2)
the mixed asymptotic regime where both variable dimension and sample size go to
infinity at comparable rates; 3) the purely high dimensional asymptotic regime
where the variable dimension goes to infinity and the sample size is fixed.
Each regime has its niche but only the latter regime applies to exa-scale data
dimension. We illustrate this high dimensional framework for the problem of
correlation mining, where it is the matrix of pairwise and partial correlations
among the variables that are of interest. We demonstrate various regimes of
correlation mining based on the unifying perspective of high dimensional
learning rates and sample complexity for different structured covariance models
and different inference tasks
Energy Consumption Of Visual Sensor Networks: Impact Of Spatio-Temporal Coverage
Wireless visual sensor networks (VSNs) are expected to play a major role in
future IEEE 802.15.4 personal area networks (PAN) under recently-established
collision-free medium access control (MAC) protocols, such as the IEEE
802.15.4e-2012 MAC. In such environments, the VSN energy consumption is
affected by the number of camera sensors deployed (spatial coverage), as well
as the number of captured video frames out of which each node processes and
transmits data (temporal coverage). In this paper, we explore this aspect for
uniformly-formed VSNs, i.e., networks comprising identical wireless visual
sensor nodes connected to a collection node via a balanced cluster-tree
topology, with each node producing independent identically-distributed
bitstream sizes after processing the video frames captured within each network
activation interval. We derive analytic results for the energy-optimal
spatio-temporal coverage parameters of such VSNs under a-priori known bounds
for the number of frames to process per sensor and the number of nodes to
deploy within each tier of the VSN. Our results are parametric to the
probability density function characterizing the bitstream size produced by each
node and the energy consumption rates of the system of interest. Experimental
results reveal that our analytic results are always within 7% of the energy
consumption measurements for a wide range of settings. In addition, results
obtained via a multimedia subsystem show that the optimal spatio-temporal
settings derived by the proposed framework allow for substantial reduction of
energy consumption in comparison to ad-hoc settings. As such, our analytic
modeling is useful for early-stage studies of possible VSN deployments under
collision-free MAC protocols prior to costly and time-consuming experiments in
the field.Comment: to appear in IEEE Transactions on Circuits and Systems for Video
Technology, 201
Internet of things
Manual of Digital Earth / Editors: Huadong Guo, Michael F. Goodchild, Alessandro Annoni .- Springer, 2020 .- ISBN: 978-981-32-9915-3Digital Earth was born with the aim of replicating the real world within the digital world. Many efforts have been made to observe and sense the Earth, both from space (remote sensing) and by using in situ sensors. Focusing on the latter, advances in Digital Earth have established vital bridges to exploit these sensors and their networks by taking location as a key element. The current era of connectivity envisions that everything is connected to everything. The concept of the Internet of Things(IoT)emergedasaholisticproposaltoenableanecosystemofvaried,heterogeneous networked objects and devices to speak to and interact with each other. To make the IoT ecosystem a reality, it is necessary to understand the electronic components, communication protocols, real-time analysis techniques, and the location of the objects and devices. The IoT ecosystem and the Digital Earth (DE) jointly form interrelated infrastructures for addressing today’s pressing issues and complex challenges. In this chapter, we explore the synergies and frictions in establishing an efficient and permanent collaboration between the two infrastructures, in order to adequately address multidisciplinary and increasingly complex real-world problems. Although there are still some pending issues, the identified synergies generate optimism for a true collaboration between the Internet of Things and the Digital Earth
Enhancing Compressed Sensing 4D Photoacoustic Tomography by Simultaneous Motion Estimation
A crucial limitation of current high-resolution 3D photoacoustic tomography
(PAT) devices that employ sequential scanning is their long acquisition time.
In previous work, we demonstrated how to use compressed sensing techniques to
improve upon this: images with good spatial resolution and contrast can be
obtained from suitably sub-sampled PAT data acquired by novel acoustic scanning
systems if sparsity-constrained image reconstruction techniques such as total
variation regularization are used. Now, we show how a further increase of image
quality can be achieved for imaging dynamic processes in living tissue (4D
PAT). The key idea is to exploit the additional temporal redundancy of the data
by coupling the previously used spatial image reconstruction models with
sparsity-constrained motion estimation models. While simulated data from a
two-dimensional numerical phantom will be used to illustrate the main
properties of this recently developed
joint-image-reconstruction-and-motion-estimation framework, measured data from
a dynamic experimental phantom will also be used to demonstrate their potential
for challenging, large-scale, real-world, three-dimensional scenarios. The
latter only becomes feasible if a carefully designed combination of tailored
optimization schemes is employed, which we describe and examine in more detail
The future of Earth observation in hydrology
In just the past 5 years, the field of Earth observation has progressed beyond the offerings of conventional space-agency-based platforms to include a plethora of sensing opportunities afforded by CubeSats, unmanned aerial vehicles (UAVs), and smartphone technologies that are being embraced by both for-profit companies and individual researchers. Over the previous decades, space agency efforts have brought forth well-known and immensely useful satellites such as the Landsat series and the Gravity Research and Climate Experiment (GRACE) system, with costs typically of the order of 1 billion dollars per satellite and with concept-to-launch timelines of the order of 2 decades (for new missions). More recently, the proliferation of smart-phones has helped to miniaturize sensors and energy requirements, facilitating advances in the use of CubeSats that can be launched by the dozens, while providing ultra-high (3-5 m) resolution sensing of the Earth on a daily basis. Start-up companies that did not exist a decade ago now operate more satellites in orbit than any space agency, and at costs that are a mere fraction of traditional satellite missions. With these advances come new space-borne measurements, such as real-time high-definition video for tracking air pollution, storm-cell development, flood propagation, precipitation monitoring, or even for constructing digital surfaces using structure-from-motion techniques. Closer to the surface, measurements from small unmanned drones and tethered balloons have mapped snow depths, floods, and estimated evaporation at sub-metre resolutions, pushing back on spatio-temporal constraints and delivering new process insights. At ground level, precipitation has been measured using signal attenuation between antennae mounted on cell phone towers, while the proliferation of mobile devices has enabled citizen scientists to catalogue photos of environmental conditions, estimate daily average temperatures from battery state, and sense other hydrologically important variables such as channel depths using commercially available wireless devices. Global internet access is being pursued via high-altitude balloons, solar planes, and hundreds of planned satellite launches, providing a means to exploit the "internet of things" as an entirely new measurement domain. Such global access will enable real-time collection of data from billions of smartphones or from remote research platforms. This future will produce petabytes of data that can only be accessed via cloud storage and will require new analytical approaches to interpret. The extent to which today's hydrologic models can usefully ingest such massive data volumes is unclear. Nor is it clear whether this deluge of data will be usefully exploited, either because the measurements are superfluous, inconsistent, not accurate enough, or simply because we lack the capacity to process and analyse them. What is apparent is that the tools and techniques afforded by this array of novel and game-changing sensing platforms present our community with a unique opportunity to develop new insights that advance fundamental aspects of the hydrological sciences. To accomplish this will require more than just an application of the technology: in some cases, it will demand a radical rethink on how we utilize and exploit these new observing systems
Nano-particle characterization by using Exposure Time Dependent Spectrum and scattering in the near field methods: how to get fast dynamics with low-speed CCD camera
Light scattering detection in the near field, a rapidly expanding family of
scattering techniques, has recently proved to be an appropriate procedure for
performing dynamic measurements. Here we report an innovative algorithm, based
on the evaluation of the Exposure Time Dependent Spectrum (ETDS), which makes
it possible to measure the fast dynamics of a colloidal suspension with the aid
of a simple near field scattering apparatus and a CCD camera. Our algorithm
consists in acquiring static spectra in the near field at different exposure
times, so that the measured decay times are limited only by the exposure time
of the camera and not by its frame rate. The experimental set-up is based on a
modified microscope, where the light scattered in the near field is collected
by a commercial objective, but (unlike in standard microscopes) the light
source is a He-Ne laser which increases the instrument sensitivity. The
apparatus and the algorithm have been validated by considering model systems of
standard spherical nano-particle
Estimating Fire Weather Indices via Semantic Reasoning over Wireless Sensor Network Data Streams
Wildfires are frequent, devastating events in Australia that regularly cause
significant loss of life and widespread property damage. Fire weather indices
are a widely-adopted method for measuring fire danger and they play a
significant role in issuing bushfire warnings and in anticipating demand for
bushfire management resources. Existing systems that calculate fire weather
indices are limited due to low spatial and temporal resolution. Localized
wireless sensor networks, on the other hand, gather continuous sensor data
measuring variables such as air temperature, relative humidity, rainfall and
wind speed at high resolutions. However, using wireless sensor networks to
estimate fire weather indices is a challenge due to data quality issues, lack
of standard data formats and lack of agreement on thresholds and methods for
calculating fire weather indices. Within the scope of this paper, we propose a
standardized approach to calculating Fire Weather Indices (a.k.a. fire danger
ratings) and overcome a number of the challenges by applying Semantic Web
Technologies to the processing of data streams from a wireless sensor network
deployed in the Springbrook region of South East Queensland. This paper
describes the underlying ontologies, the semantic reasoning and the Semantic
Fire Weather Index (SFWI) system that we have developed to enable domain
experts to specify and adapt rules for calculating Fire Weather Indices. We
also describe the Web-based mapping interface that we have developed, that
enables users to improve their understanding of how fire weather indices vary
over time within a particular region.Finally, we discuss our evaluation results
that indicate that the proposed system outperforms state-of-the-art techniques
in terms of accuracy, precision and query performance.Comment: 20pages, 12 figure
Quality of Information in Mobile Crowdsensing: Survey and Research Challenges
Smartphones have become the most pervasive devices in people's lives, and are
clearly transforming the way we live and perceive technology. Today's
smartphones benefit from almost ubiquitous Internet connectivity and come
equipped with a plethora of inexpensive yet powerful embedded sensors, such as
accelerometer, gyroscope, microphone, and camera. This unique combination has
enabled revolutionary applications based on the mobile crowdsensing paradigm,
such as real-time road traffic monitoring, air and noise pollution, crime
control, and wildlife monitoring, just to name a few. Differently from prior
sensing paradigms, humans are now the primary actors of the sensing process,
since they become fundamental in retrieving reliable and up-to-date information
about the event being monitored. As humans may behave unreliably or
maliciously, assessing and guaranteeing Quality of Information (QoI) becomes
more important than ever. In this paper, we provide a new framework for
defining and enforcing the QoI in mobile crowdsensing, and analyze in depth the
current state-of-the-art on the topic. We also outline novel research
challenges, along with possible directions of future work.Comment: To appear in ACM Transactions on Sensor Networks (TOSN
- …