13,808 research outputs found

    Gridless Two-dimensional DOA Estimation With L-shaped Array Based on the Cross-covariance Matrix

    Full text link
    The atomic norm minimization (ANM) has been successfully incorporated into the two-dimensional (2-D) direction-of-arrival (DOA) estimation problem for super-resolution. However, its computational workload might be unaffordable when the number of snapshots is large. In this paper, we propose two gridless methods for 2-D DOA estimation with L-shaped array based on the atomic norm to improve the computational efficiency. Firstly, by exploiting the cross-covariance matrix an ANM-based model has been proposed. We then prove that this model can be efficiently solved as a semi-definite programming (SDP). Secondly, a modified model has been presented to improve the estimation accuracy. It is shown that our proposed methods can be applied to both uniform and sparse L-shaped arrays and do not require any knowledge of the number of sources. Furthermore, since our methods greatly reduce the model size as compared to the conventional ANM method, and thus are much more efficient. Simulations results are provided to demonstrate the advantage of our methods

    CanvasGAN: A simple baseline for text to image generation by incrementally patching a canvas

    Full text link
    We propose a new recurrent generative model for generating images from text captions while attending on specific parts of text captions. Our model creates images by incrementally adding patches on a "canvas" while attending on words from text caption at each timestep. Finally, the canvas is passed through an upscaling network to generate images. We also introduce a new method for generating visual-semantic sentence embeddings based on self-attention over text. We compare our model's generated images with those generated Reed et. al.'s model and show that our model is a stronger baseline for text to image generation tasks.Comment: CVC 201

    Image-to-Image Translation with Conditional Adversarial Networks

    Full text link
    We investigate conditional adversarial networks as a general-purpose solution to image-to-image translation problems. These networks not only learn the mapping from input image to output image, but also learn a loss function to train this mapping. This makes it possible to apply the same generic approach to problems that traditionally would require very different loss formulations. We demonstrate that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks. Indeed, since the release of the pix2pix software associated with this paper, a large number of internet users (many of them artists) have posted their own experiments with our system, further demonstrating its wide applicability and ease of adoption without the need for parameter tweaking. As a community, we no longer hand-engineer our mapping functions, and this work suggests we can achieve reasonable results without hand-engineering our loss functions either.Comment: Website: https://phillipi.github.io/pix2pix/, CVPR 201
    corecore