9,706 research outputs found
Disparities in Cause-Specific Cancer Survival by Census Tract Poverty Level in Idaho, U.S.
Objective. This population-based study compared cause-specific cancer survival by socioeconomic status using methods to more accurately assign cancer deaths to primary site. Methods. The current study analyzed Idaho data used in the Accuracy of Cancer Mortality Statistics Based on Death Certificates (ACM) study supplemented with additional information to measure cause-specific cancer survival by census tract poverty level. Results. The distribution of cases by primary site group differed significantly by poverty level (chi-square = 265.3, 100 df, p In the life table analyses, for 8 of 24 primary site groups investigated, and all sites combined, there was a significant gradient relating higher poverty with poorer survival. For all sites combined, the absolute difference in 5-year cause-specific survival rate was 13.6% between the lowest and highest poverty levels. Conclusions. This study shows striking disparities in cause-specific cancer survival related to the poverty level of the area a person resides in at the time of diagnosis
Focus Is All You Need: Loss Functions For Event-based Vision
Event cameras are novel vision sensors that output pixel-level brightness
changes ("events") instead of traditional video frames. These asynchronous
sensors offer several advantages over traditional cameras, such as, high
temporal resolution, very high dynamic range, and no motion blur. To unlock the
potential of such sensors, motion compensation methods have been recently
proposed. We present a collection and taxonomy of twenty two objective
functions to analyze event alignment in motion compensation approaches (Fig.
1). We call them Focus Loss Functions since they have strong connections with
functions used in traditional shape-from-focus applications. The proposed loss
functions allow bringing mature computer vision tools to the realm of event
cameras. We compare the accuracy and runtime performance of all loss functions
on a publicly available dataset, and conclude that the variance, the gradient
and the Laplacian magnitudes are among the best loss functions. The
applicability of the loss functions is shown on multiple tasks: rotational
motion, depth and optical flow estimation. The proposed focus loss functions
allow to unlock the outstanding properties of event cameras.Comment: 29 pages, 19 figures, 4 table
- âŠ