1,633 research outputs found
A single-photon sampling architecture for solid-state imaging
Advances in solid-state technology have enabled the development of silicon
photomultiplier sensor arrays capable of sensing individual photons. Combined
with high-frequency time-to-digital converters (TDCs), this technology opens up
the prospect of sensors capable of recording with high accuracy both the time
and location of each detected photon. Such a capability could lead to
significant improvements in imaging accuracy, especially for applications
operating with low photon fluxes such as LiDAR and positron emission
tomography.
The demands placed on on-chip readout circuitry imposes stringent trade-offs
between fill factor and spatio-temporal resolution, causing many contemporary
designs to severely underutilize the technology's full potential. Concentrating
on the low photon flux setting, this paper leverages results from group testing
and proposes an architecture for a highly efficient readout of pixels using
only a small number of TDCs, thereby also reducing both cost and power
consumption. The design relies on a multiplexing technique based on binary
interconnection matrices. We provide optimized instances of these matrices for
various sensor parameters and give explicit upper and lower bounds on the
number of TDCs required to uniquely decode a given maximum number of
simultaneous photon arrivals.
To illustrate the strength of the proposed architecture, we note a typical
digitization result of a 120x120 photodiode sensor on a 30um x 30um pitch with
a 40ps time resolution and an estimated fill factor of approximately 70%, using
only 161 TDCs. The design guarantees registration and unique recovery of up to
4 simultaneous photon arrivals using a fast decoding algorithm. In a series of
realistic simulations of scintillation events in clinical positron emission
tomography the design was able to recover the spatio-temporal location of 98.6%
of all photons that caused pixel firings.Comment: 24 pages, 3 figures, 5 table
Amorphous Placement and Informed Diffusion for Timely Monitoring by Autonomous, Resource-Constrained, Mobile Sensors
Personal communication devices are increasingly equipped with sensors for passive monitoring of encounters and surroundings. We envision the emergence of services that enable a community of mobile users carrying such resource-limited devices to query such information at remote locations in the field in which they collectively roam. One approach to implement such a service is directed placement and retrieval (DPR), whereby readings/queries about a specific location are routed to a node responsible for that location. In a mobile, potentially sparse setting, where end-to-end paths are unavailable, DPR is not an attractive solution as it would require the use of delay-tolerant (flooding-based store-carry-forward) routing of both readings and queries, which is inappropriate for applications with data freshness constraints, and which is incompatible with stringent device power/memory constraints. Alternatively, we propose the use of amorphous placement and retrieval (APR), in which routing and field monitoring are integrated through the use of a cache management scheme coupled with an informed exchange of cached samples to diffuse sensory data throughout the network, in such a way that a query answer is likely to be found close to the query origin. We argue that knowledge of the distribution of query targets could be used effectively by an informed cache management policy to maximize the utility of collective storage of all devices. Using a simple analytical model, we show that the use of informed cache management is particularly important when the mobility model results in a non-uniform distribution of users over the field. We present results from extensive simulations which show that in sparsely-connected networks, APR is more cost-effective than DPR, that it provides extra resilience to node failure and packet losses, and that its use of informed cache management yields superior performance
GROWTH on S190510g: DECam Observation Planning and Follow-Up of a Distant Binary Neutron Star Merger Candidate
The first two months of the third Advanced LIGO and Virgo observing run (2019 April–May) showed that distant gravitational-wave (GW) events can now be readily detected. Three candidate mergers containing neutron stars (NS) were reported in a span of 15 days, all likely located more than 100 Mpc away. However, distant events such as the three new NS mergers are likely to be coarsely localized, which highlights the importance of facilities and scheduling systems that enable deep observations over hundreds to thousands of square degrees to detect the electromagnetic counterparts. On 2019 May 10 02:59:39.292 UT the GW candidate S190510g was discovered and initially classified as a binary neutron star (BNS) merger with 98% probability. The GW event was localized within an area of 3462 deg^2, later refined to 1166 deg^2 (90%) at a distance of 227 ± 92 Mpc. We triggered Target-of-Opportunity observations with the Dark Energy Camera (DECam), a wide-field optical imager mounted at the prime focus of the 4 m Blanco Telescope at Cerro Tololo Inter-American Observatory in Chile. This Letter describes our DECam observations and our real-time analysis results, focusing in particular on the design and implementation of the observing strategy. Within 24 hr of the merger time, we observed 65% of the total enclosed probability of the final skymap with an observing efficiency of 94%. We identified and publicly announced 13 candidate counterparts. S190510g was reclassified 1.7 days after the merger, after our observations were completed, with a "BNS merger" probability reduced from 98% to 42% in favor of a "terrestrial classification
Finding events in temporal networks : segmentation meets densest subgraph discovery
In this paper, we study the problem of discovering a timeline of events in a temporal network. We model events as dense subgraphs that occur within intervals of network activity. We formulate the event discovery task as an optimization problem, where we search for a partition of the network timeline into k non-overlapping intervals, such that the intervals span subgraphs with maximum total density. The output is a sequence of dense subgraphs along with corresponding time intervals, capturing the most interesting events during the network lifetime. A naïve solution to our optimization problem has polynomial but prohibitively high running time. We adapt existing recent work on dynamic densest subgraph discovery and approximate dynamic programming to design a fast approximation algorithm. Next, to ensure richer structure, we adjust the problem formulation to encourage coverage of a larger set of nodes. This problem is NP-hard; however, we show that on static graphs a simple greedy algorithm leads to approximate solution due to submodularity. We extend this greedy approach for temporal networks, but we lose the approximation guarantee in the process. Finally, we demonstrate empirically that our algorithms recover solutions with good quality.Peer reviewe
Quantum algorithms applied to satellite mission planning for Earth observation
Earth imaging satellites are a crucial part of our everyday lives that enable
global tracking of industrial activities. Use cases span many applications,
from weather forecasting to digital maps, carbon footprint tracking, and
vegetation monitoring. However, there are also limitations; satellites are
difficult to manufacture, expensive to maintain, and tricky to launch into
orbit. Therefore, it is critical that satellites are employed efficiently. This
poses a challenge known as the satellite mission planning problem, which could
be computationally prohibitive to solve on large scales. However,
close-to-optimal algorithms can often provide satisfactory resolutions, such as
greedy reinforcement learning, and optimization algorithms. This paper
introduces a set of quantum algorithms to solve the mission planning problem
and demonstrate an advantage over the classical algorithms implemented thus
far. The problem is formulated as maximizing the number of high-priority tasks
completed on real datasets containing thousands of tasks and multiple
satellites. This work demonstrates that through solution-chaining and
clustering, optimization and machine learning algorithms offer the greatest
potential for optimal solutions. Most notably, this paper illustrates that a
hybridized quantum-enhanced reinforcement learning agent can achieve a
completion percentage of 98.5% over high-priority tasks, which is a significant
improvement over the baseline greedy methods with a completion rate of 63.6%.
The results presented in this work pave the way to quantum-enabled solutions in
the space industry and, more generally, future mission planning problems across
industries.Comment: 13 pages, 10 figues, 3 tables. Submitted to IEEE JSTAR
- …