402 research outputs found

    Bi-Objective Nonnegative Matrix Factorization: Linear Versus Kernel-Based Models

    Full text link
    Nonnegative matrix factorization (NMF) is a powerful class of feature extraction techniques that has been successfully applied in many fields, namely in signal and image processing. Current NMF techniques have been limited to a single-objective problem in either its linear or nonlinear kernel-based formulation. In this paper, we propose to revisit the NMF as a multi-objective problem, in particular a bi-objective one, where the objective functions defined in both input and feature spaces are taken into account. By taking the advantage of the sum-weighted method from the literature of multi-objective optimization, the proposed bi-objective NMF determines a set of nondominated, Pareto optimal, solutions instead of a single optimal decomposition. Moreover, the corresponding Pareto front is studied and approximated. Experimental results on unmixing real hyperspectral images confirm the efficiency of the proposed bi-objective NMF compared with the state-of-the-art methods

    A Method for Finding Structured Sparse Solutions to Non-negative Least Squares Problems with Applications

    Full text link
    Demixing problems in many areas such as hyperspectral imaging and differential optical absorption spectroscopy (DOAS) often require finding sparse nonnegative linear combinations of dictionary elements that match observed data. We show how aspects of these problems, such as misalignment of DOAS references and uncertainty in hyperspectral endmembers, can be modeled by expanding the dictionary with grouped elements and imposing a structured sparsity assumption that the combinations within each group should be sparse or even 1-sparse. If the dictionary is highly coherent, it is difficult to obtain good solutions using convex or greedy methods, such as non-negative least squares (NNLS) or orthogonal matching pursuit. We use penalties related to the Hoyer measure, which is the ratio of the l1l_1 and l2l_2 norms, as sparsity penalties to be added to the objective in NNLS-type models. For solving the resulting nonconvex models, we propose a scaled gradient projection algorithm that requires solving a sequence of strongly convex quadratic programs. We discuss its close connections to convex splitting methods and difference of convex programming. We also present promising numerical results for example DOAS analysis and hyperspectral demixing problems.Comment: 38 pages, 14 figure

    Implementation strategies for hyperspectral unmixing using Bayesian source separation

    Get PDF
    Bayesian Positive Source Separation (BPSS) is a useful unsupervised approach for hyperspectral data unmixing, where numerical non-negativity of spectra and abundances has to be ensured, such in remote sensing. Moreover, it is sensible to impose a sum-to-one (full additivity) constraint to the estimated source abundances in each pixel. Even though non-negativity and full additivity are two necessary properties to get physically interpretable results, the use of BPSS algorithms has been so far limited by high computation time and large memory requirements due to the Markov chain Monte Carlo calculations. An implementation strategy which allows one to apply these algorithms on a full hyperspectral image, as typical in Earth and Planetary Science, is introduced. Effects of pixel selection, the impact of such sampling on the relevance of the estimated component spectra and abundance maps, as well as on the computation times, are discussed. For that purpose, two different dataset have been used: a synthetic one and a real hyperspectral image from Mars.Comment: 10 pages, 6 figures, submitted to IEEE Transactions on Geoscience and Remote Sensing in the special issue on Hyperspectral Image and Signal Processing (WHISPERS
    corecore