57,463 research outputs found

    A review of multi-instance learning assumptions

    Get PDF
    Multi-instance (MI) learning is a variant of inductive machine learning, where each learning example contains a bag of instances instead of a single feature vector. The term commonly refers to the supervised setting, where each bag is associated with a label. This type of representation is a natural fit for a number of real-world learning scenarios, including drug activity prediction and image classification, hence many MI learning algorithms have been proposed. Any MI learning method must relate instances to bag-level class labels, but many types of relationships between instances and class labels are possible. Although all early work in MI learning assumes a specific MI concept class known to be appropriate for a drug activity prediction domain; this ‘standard MI assumption’ is not guaranteed to hold in other domains. Much of the recent work in MI learning has concentrated on a relaxed view of the MI problem, where the standard MI assumption is dropped, and alternative assumptions are considered instead. However, often it is not clearly stated what particular assumption is used and how it relates to other assumptions that have been proposed. In this paper, we aim to clarify the use of alternative MI assumptions by reviewing the work done in this area

    High-order integral equation methods for problems of scattering by bumps and cavities on half-planes

    Get PDF
    This paper presents high-order integral equation methods for evaluation of electromagnetic wave scattering by dielectric bumps and dielectric cavities on perfectly conducting or dielectric half-planes. In detail, the algorithms introduced in this paper apply to eight classical scattering problems, namely: scattering by a dielectric bump on a perfectly conducting or a dielectric half-plane, and scattering by a filled, overfilled or void dielectric cavity on a perfectly conducting or a dielectric half-plane. In all cases field representations based on single-layer potentials for appropriately chosen Green functions are used. The numerical far fields and near fields exhibit excellent convergence as discretizations are refined--even at and around points where singular fields and infinite currents exist.Comment: 25 pages, 7 figure

    Unit commitment with valve-point loading effect

    Full text link
    Valve-point loading affects the input-output characteristics of generating units, bringing the fuel costs nonlinear and nonsmooth. This has been considered in the solution of load dispatch problems, but not in the planning phase of unit commitment. This paper presents a mathematical optimization model for the thermal unit commitment problem considering valve-point loading. The formulation is based on a careful linearization of the fuel cost function, which is modeled with great detail on power regions being used in the current solution, and roughly on other regions. A set of benchmark instances for this problem is used for analyzing the method, with recourse to a general-purpose mixed-integer optimization solver

    On Precoding for Constant K-User MIMO Gaussian Interference Channel with Finite Constellation Inputs

    Full text link
    This paper considers linear precoding for constant channel-coefficient KK-User MIMO Gaussian Interference Channel (MIMO GIC) where each transmitter-ii (Tx-ii), requires to send did_i independent complex symbols per channel use that take values from fixed finite constellations with uniform distribution, to receiver-ii (Rx-ii) for i=1,2,,Ki=1,2,\cdots,K. We define the maximum rate achieved by Tx-ii using any linear precoder, when the interference channel-coefficients are zero, as the signal to noise ratio (SNR) tends to infinity to be the Constellation Constrained Saturation Capacity (CCSC) for Tx-ii. We derive a high SNR approximation for the rate achieved by Tx-ii when interference is treated as noise and this rate is given by the mutual information between Tx-ii and Rx-ii, denoted as I[Xi;Yi]I[X_i;Y_i]. A set of necessary and sufficient conditions on the precoders under which I[Xi;Yi]I[X_i;Y_i] tends to CCSC for Tx-ii is derived. Interestingly, the precoders designed for interference alignment (IA) satisfy these necessary and sufficient conditions. Further, we propose gradient-ascent based algorithms to optimize the sum-rate achieved by precoding with finite constellation inputs and treating interference as noise. Simulation study using the proposed algorithms for a 3-user MIMO GIC with two antennas at each node with di=1d_i=1 for all ii, and with BPSK and QPSK inputs, show more than 0.1 bits/sec/Hz gain in the ergodic sum-rate over that yielded by precoders obtained from some known IA algorithms, at moderate SNRs.Comment: 15 pages, 9 figure

    Sampling-based optimal kinodynamic planning with motion primitives

    Full text link
    This paper proposes a novel sampling-based motion planner, which integrates in RRT* (Rapidly exploring Random Tree star) a database of pre-computed motion primitives to alleviate its computational load and allow for motion planning in a dynamic or partially known environment. The database is built by considering a set of initial and final state pairs in some grid space, and determining for each pair an optimal trajectory that is compatible with the system dynamics and constraints, while minimizing a cost. Nodes are progressively added to the tree {of feasible trajectories in the RRT* by extracting at random a sample in the gridded state space and selecting the best obstacle-free motion primitive in the database that joins it to an existing node. The tree is rewired if some nodes can be reached from the new sampled state through an obstacle-free motion primitive with lower cost. The computationally more intensive part of motion planning is thus moved to the preliminary offline phase of the database construction at the price of some performance degradation due to gridding. Grid resolution can be tuned so as to compromise between (sub)optimality and size of the database. The planner is shown to be asymptotically optimal as the grid resolution goes to zero and the number of sampled states grows to infinity

    Hierarchical structure-and-motion recovery from uncalibrated images

    Full text link
    This paper addresses the structure-and-motion problem, that requires to find camera motion and 3D struc- ture from point matches. A new pipeline, dubbed Samantha, is presented, that departs from the prevailing sequential paradigm and embraces instead a hierarchical approach. This method has several advantages, like a provably lower computational complexity, which is necessary to achieve true scalability, and better error containment, leading to more stability and less drift. Moreover, a practical autocalibration procedure allows to process images without ancillary information. Experiments with real data assess the accuracy and the computational efficiency of the method.Comment: Accepted for publication in CVI

    Foundational principles for large scale inference: Illustrations through correlation mining

    Full text link
    When can reliable inference be drawn in the "Big Data" context? This paper presents a framework for answering this fundamental question in the context of correlation mining, with implications for general large scale inference. In large scale data applications like genomics, connectomics, and eco-informatics the dataset is often variable-rich but sample-starved: a regime where the number nn of acquired samples (statistical replicates) is far fewer than the number pp of observed variables (genes, neurons, voxels, or chemical constituents). Much of recent work has focused on understanding the computational complexity of proposed methods for "Big Data." Sample complexity however has received relatively less attention, especially in the setting when the sample size nn is fixed, and the dimension pp grows without bound. To address this gap, we develop a unified statistical framework that explicitly quantifies the sample complexity of various inferential tasks. Sampling regimes can be divided into several categories: 1) the classical asymptotic regime where the variable dimension is fixed and the sample size goes to infinity; 2) the mixed asymptotic regime where both variable dimension and sample size go to infinity at comparable rates; 3) the purely high dimensional asymptotic regime where the variable dimension goes to infinity and the sample size is fixed. Each regime has its niche but only the latter regime applies to exa-scale data dimension. We illustrate this high dimensional framework for the problem of correlation mining, where it is the matrix of pairwise and partial correlations among the variables that are of interest. We demonstrate various regimes of correlation mining based on the unifying perspective of high dimensional learning rates and sample complexity for different structured covariance models and different inference tasks
    corecore