338 research outputs found

    Model Averaging by Cross-validation for Partially Linear Functional Additive Models

    Full text link
    In this paper, we propose a model averaging approach for addressing model uncertainty in the context of partial linear functional additive models. These models are designed to describe the relation between a response and mixed-types of predictors by incorporating both the parametric effect of scalar variables and the additive effect of a functional variable. The proposed model averaging scheme assigns weights to candidate models based on the minimization of a multi-fold cross-validation criterion. Furthermore, we establish the asymptotic optimality of the resulting estimator in terms of achieving the lowest possible square prediction error loss under model misspecification. Extensive simulation studies and an application to a near infrared spectra dataset are presented to support and illustrate our method

    A hierarchical semantic segmentation framework for computer vision-based bridge damage detection

    Full text link
    Computer vision-based damage detection using remote cameras and unmanned aerial vehicles (UAVs) enables efficient and low-cost bridge health monitoring that reduces labor costs and the needs for sensor installation and maintenance. By leveraging recent semantic image segmentation approaches, we are able to find regions of critical structural components and recognize damage at the pixel level using images as the only input. However, existing methods perform poorly when detecting small damages (e.g., cracks and exposed rebars) and thin objects with limited image samples, especially when the components of interest are highly imbalanced. To this end, this paper introduces a semantic segmentation framework that imposes the hierarchical semantic relationship between component category and damage types. For example, certain concrete cracks only present on bridge columns and therefore the non-column region will be masked out when detecting such damages. In this way, the damage detection model could focus on learning features from possible damaged regions only and avoid the effects of other irrelevant regions. We also utilize multi-scale augmentation that provides views with different scales that preserves contextual information of each image without losing the ability of handling small and thin objects. Furthermore, the proposed framework employs important sampling that repeatedly samples images containing rare components (e.g., railway sleeper and exposed rebars) to provide more data samples, which addresses the imbalanced data challenge

    Estimating optimal treatment regimes in survival contexts using an instrumental variable

    Full text link
    In survival contexts, substantial literature exists on estimating optimal treatment regimes, where treatments are assigned based on personal characteristics for the purpose of maximizing the survival probability. These methods assume that a set of covariates is sufficient to deconfound the treatment-outcome relationship. Nevertheless, the assumption can be limiting in observational studies or randomized trials in which noncompliance occurs. Thus, we advance a novel approach for estimating the optimal treatment regime when certain confounders are not observable and a binary instrumental variable is available. Specifically, via a binary instrumental variable, we propose two semiparametric estimators for the optimal treatment regime, one of which possesses the desirable property of double robustness, by maximizing Kaplan-Meier-like estimators within a pre-defined class of regimes. Because the Kaplan-Meier-like estimators are jagged, we incorporate kernel smoothing methods to enhance their performance. Under appropriate regularity conditions, the asymptotic properties are rigorously established. Furthermore, the finite sample performance is assessed through simulation studies. We exemplify our method using data from the National Cancer Institute's (NCI) prostate, lung, colorectal, and ovarian cancer screening trial

    Adaptive Beamforming for Distributed Relay Networks

    Get PDF
    Tremendous research work has been put into the realm of distributed relay networks, for its distinct advantages in exploiting spatial diversity, reducing the deployment cost and mitigating the effect of fading in wireless transmission without the multi-antenna requirement on the relay nodes. In typical relay networks, data transmission between a source and a destination is assisted by relay nodes with various relaying protocols. In this thesis, we investigate how to adaptively select the relay weights to meet specific interference suppressing requirements of the network. The thesis makes original contributions by proposing a filter-and-forward (FF) relay scheme in cognitive radio networks and an iterative algorithm based transceiver beamforming scheme for multi-pair relay networks. In the firstly proposed scheme, the relay nodes are adapted to deal with the inter-symbol-interference (ISI) that is introduced in the frequency-selective channel environment and the leakage interference introduced to the primary user. Our proposed scheme uses FF relay beamforming at the relay nodes to combat the frequency selective channel, and our scheme also aims to maximize the received SINR at the secondary destination, while suppressing the interference introduced to the primary user (PU). This scheme is further extended to accommodate a relay nodes output power constraint. Under certain criteria, the extended scheme can be transformed into two sub-schemes with lower computational complexity, where their closed-form solutions are derived. The probability that we can perform these transformations is also tested, which reveals under what circumstances our second scheme can be solved more easily. Then, we propose an iterative transceiver beamforming scheme for the multi-pair distributed relay networks. In our scheme, we consider multi-antenna users in one user group communicating with their partners in the other user group via distributed single-antenna relay nodes. We employ transceiver beamformers at the user nodes, and through our proposed iterative algorithm the relay nodes and user nodes can be coordinatively adapted to suppress the inter-pair-interference (IPI) while maximize the desired signal power. We also divide the rather difficult transceiver beamforming problem into three sub-problems, each of which can be solved with sub-optimal solutions. The transmit beamforming vectors, distributed relay coefficients and the receive beamforming vectors are obtained by iteratively solving these three sub-problems, each having a closed-form solution. The tasks of maximizing desired signal power, and reducing inter-pair interference (IPI) and noise are thus allocated to different iteration steps. By this arrangement, the transmit and receiver beamformers of each user are responsible for improving its own performance and the distributed relay nodes can be employed with simple amplify-and-forward(AF) protocols and only forward the received signal with proper scalar. This iterative relay beamforming scheme is further extended by distributing the computation tasks among each user and relay node, through which high computational efficiency can be ensured while extra overhead of bandwidth is need for sharing beamforming vector updates during the iteration steps. Furthermore, with respect to the channel uncertainty, two more relay strategies are proposed considering two different requirements from the communication network: sum relay output power and individual relay output power. At last, the application of the iterative relay beamforming method in cognitive radio networks is studied, where multiple pairs of users are considered as secondary users (SUs), and the designed transmit beamforming vector, relay beamforming vector and receive beamforming vector together guarantee that the inner interference of their transmissions is well suppressed while the interference introduced by them to the PU is restricted under a predefined threshold

    Augmented Deep Representations for Unconstrained Still/Video-based Face Recognition

    Get PDF
    Face recognition is one of the active areas of research in computer vision and biometrics. Many approaches have been proposed in the literature that demonstrate impressive performance, especially those based on deep learning. However, unconstrained face recognition with large pose, illumination, occlusion and other variations is still an unsolved problem. Unconstrained video-based face recognition is even more challenging due to the large volume of data to be processed, lack of labeled training data and significant intra/inter-video variations on scene, blur, video quality, etc. Although Deep Convolutional Neural Networks (DCNNs) have provided discriminant representations for faces and achieved performance surpassing humans in controlled scenarios, modifications are necessary for face recognition in unconstrained conditions. In this dissertation, we propose several methods that improve unconstrained face recognition performance by augmenting the representation provided by the deep networks using correlation or contextual information in the data. For unconstrained still face recognition, we present an encoding approach to combine the Fisher vector (FV) encoding and DCNN representations, which is called FV-DCNN. The feature maps from the last convolutional layer in the deep network are encoded by FV into a robust representation, which utilizes the correlation between facial parts within each face. A VLAD-based encoding method called VLAD-DCNN is also proposed as an extension. Extensive evaluations on three challenging face recognition datasets show that the proposed FV-DCNN and VLAD-DCNN perform comparable to or better than many state-of-the-art face verification methods. For the more challenging video-based face recognition task, we first propose an automatic system and model the video-to-video similarity as subspace-to-subspace similarity, where the subspaces characterize the correlation between deep representations of faces in videos. In the system, a quality-aware subspace-to-subspace similarity is introduced, where subspaces are learned using quality-aware principal component analysis. Subspaces along with quality-aware exemplars of templates are used to produce the similarity scores between video pairs by a quality-aware principal angle-based subspace-to-subspace similarity metric. The method is evaluated on four video datasets. The experimental results demonstrate the superior performance of the proposed method. To utilize the temporal information in videos, a hybrid dictionary learning method is also proposed for video-based face recognition. The proposed unsupervised approach effectively models the temporal correlation between deep representations of video faces using dynamical dictionaries. A practical iterative optimization algorithm is introduced to learn the dynamical dictionary. Experiments on three video-based face recognition datasets demonstrate that the proposed method can effectively learn robust and discriminative representation for videos and improve the face recognition performance. Finally, to leverage contextual information in videos, we present the Uncertainty-Gated Graph (UGG) for unconstrained video-based face recognition. It utilizes contextual information between faces by conducting graph-based identity propagation between sample tracklets, where identity information are initialized by the deep representations of video faces. UGG explicitly models the uncertainty of the contextual connections between tracklets by adaptively updating the weights of the edge gates according to the identity distributions of the nodes during inference. UGG is a generic graphical model that can be applied at only inference time or with end-to-end training. We demonstrate the effectiveness of UGG with state-of-the-art results on the recently released challenging Cast Search in Movies and IARPA Janus Surveillance Video Benchmark datasets
    • …
    corecore