401,090 research outputs found

    Improving Facial Attribute Prediction using Semantic Segmentation

    Full text link
    Attributes are semantically meaningful characteristics whose applicability widely crosses category boundaries. They are particularly important in describing and recognizing concepts where no explicit training example is given, \textit{e.g., zero-shot learning}. Additionally, since attributes are human describable, they can be used for efficient human-computer interaction. In this paper, we propose to employ semantic segmentation to improve facial attribute prediction. The core idea lies in the fact that many facial attributes describe local properties. In other words, the probability of an attribute to appear in a face image is far from being uniform in the spatial domain. We build our facial attribute prediction model jointly with a deep semantic segmentation network. This harnesses the localization cues learned by the semantic segmentation to guide the attention of the attribute prediction to the regions where different attributes naturally show up. As a result of this approach, in addition to recognition, we are able to localize the attributes, despite merely having access to image level labels (weak supervision) during training. We evaluate our proposed method on CelebA and LFWA datasets and achieve superior results to the prior arts. Furthermore, we show that in the reverse problem, semantic face parsing improves when facial attributes are available. That reaffirms the need to jointly model these two interconnected tasks

    Object detection via a multi-region & semantic segmentation-aware CNN model

    Get PDF
    We propose an object detection system that relies on a multi-region deep convolutional neural network (CNN) that also encodes semantic segmentation-aware features. The resulting CNN-based representation aims at capturing a diverse set of discriminative appearance factors and exhibits localization sensitivity that is essential for accurate object localization. We exploit the above properties of our recognition module by integrating it on an iterative localization mechanism that alternates between scoring a box proposal and refining its location with a deep CNN regression model. Thanks to the efficient use of our modules, we detect objects with very high localization accuracy. On the detection challenges of PASCAL VOC2007 and PASCAL VOC2012 we achieve mAP of 78.2% and 73.9% correspondingly, surpassing any other published work by a significant margin.Comment: Extended technical report -- short version to appear at ICCV 201

    Can Negligible Cooperation Increase Network Reliability?

    Get PDF
    In network cooperation strategies, nodes work together with the aim of increasing transmission rates or reliability. This paper demonstrates that enabling cooperation between the transmitters of a two-user multiple access channel, via a cooperation facilitator that has access to both messages, always results in a network whose maximal- and average-error sum-capacities are the same---even when those capacities differ in the absence of cooperation and the information shared with the encoders is negligible. From this result, it follows that if a multiple access channel with no transmitter cooperation has different maximal- and average-error sum-capacities, then the maximal-error sum-capacity of the network consisting of this channel and a cooperation facilitator is not continuous with respect to the output edge capacities of the facilitator. This shows that there exist networks where sharing even a negligible number of bits per channel use with the encoders yields a non-negligible benefit.Comment: 27 pages, 3 figures. Submitted to the IEEE Transactions on Information Theor

    Robust Linear Precoder Design for Multi-cell Downlink Transmission

    Full text link
    Coordinated information processing by the base stations of multi-cell wireless networks enhances the overall quality of communication in the network. Such coordinations for optimizing any desired network-wide quality of service (QoS) necessitate the base stations to acquire and share some channel state information (CSI). With perfect knowledge of channel states, the base stations can adjust their transmissions for achieving a network-wise QoS optimality. In practice, however, the CSI can be obtained only imperfectly. As a result, due to the uncertainties involved, the network is not guaranteed to benefit from a globally optimal QoS. Nevertheless, if the channel estimation perturbations are confined within bounded regions, the QoS measure will also lie within a bounded region. Therefore, by exploiting the notion of robustness in the worst-case sense some worst-case QoS guarantees for the network can be asserted. We adopt a popular model for noisy channel estimates that assumes that estimation noise terms lie within known hyper-spheres. We aim to design linear transceivers that optimize a worst-case QoS measure in downlink transmissions. In particular, we focus on maximizing the worst-case weighted sum-rate of the network and the minimum worst-case rate of the network. For obtaining such transceiver designs, we offer several centralized (fully cooperative) and distributed (limited cooperation) algorithms which entail different levels of complexity and information exchange among the base stations.Comment: 38 Pages, 7 Figures, To appear in the IEEE Transactions on Signal Processin
    • …
    corecore