337 research outputs found

    Iterated maxima search for the maximally diverse grouping problem

    Get PDF
    The maximally diverse grouping problem (MDGP) is to partition the vertices of an edge-weighted and undirected complete graph into m groups such that the total weight of the groups is maximized subject to some group size constraints. MDGP is a NP-hard combinatorial problem with a number of relevant applications. In this paper, we present an innovative heuristic algorithm called iterated maxima search (IMS) algorithm for solving MDGP. The proposed approach employs a maxima search procedure that integrates organically an efficient local optimization method and a weak perturbation operator to reinforce the intensification of the search and a strong perturbation operator to diversify the search. Extensive experiments on five sets of 500 MDGP benchmark instances of the literature show that IMS competes favorably with the state-of-the-art algorithms. We provide additional experiments to shed light on the rationality of the proposed algorithm and investigate the role of the key ingredients

    An iterated greedy heuristic for a market segmentation problem with multiple attributes

    Full text link
    [EN] A real-world customer segmentation problem from a beverage distribution firm is addressed. The firm wants to partition a set of customers, who share geographical and marketing attributes, into segments according to certain requirements: (a) customers allocated to the same segment must have very similar attributes: type of contract, type of store and the average difference of purchase volume; and (b) compact segments are desired. The main reason for creating a partition with these features is because the firm wants to try different product marketing strategies. In this paper, a detailed attribute formulation and an iterated greedy heuristic that iteratively destroys and reconstructs a given partition are proposed. The initial partition is obtained by using a modified k-means algorithm that involves a GRASP philosophy to get the initial configuration of centers. The heuristic includes an improvement method that employs two local search procedures. Computational results and statistical analyses show the effectiveness of the proposed approach and its individual components. The proposed metaheuristic is also observed very competitive, faster, and more robust when compared to existing methods. (C) 2017 Elsevier B.V. All rights reserved.This research has been supported by the Mexican National Council for Science and Technology (CONACYT) through grants CB2005-01-48499Y and CB2011-01-166397, and a scholarship for graduate studies, and by the Universidad Autonoma de Nuevo Leon through its Scientific and Technological Research Support Program (PAICYT), grants CA1478-07, CE012-09, IT511-10, and CE331-15. Ruben Ruiz is partially supported by the Spanish Ministry of Economy and Competitiveness, under the project "SCHEYARD - Optimization of Scheduling Problems in Container Yards" (No. DPI2015-65895-R) financed by FEDER funds. We would like to thank Rafael Frinhani, Richard Fuchshuber, and their corresponding research teams for providing us the source code of their algorithms to carry out the corresponding tests. Furthermore, we are grateful to the editor and the four anonymous reviewers for their careful reading of our manuscript and their constructive comments and suggestions which helped us improve its quality.Huerta-Muñoz, D.; Ríos-Mercado, RZ.; Ruiz García, R. (2017). An iterated greedy heuristic for a market segmentation problem with multiple attributes. European Journal of Operational Research. 261(1):75-87. https://doi.org/10.1016/j.ejor.2017.02.013S7587261

    Score-Based Approaches to Heterogeneity in Psychological Models

    Get PDF
    Statistische Modelle menschlicher Kognition und Verhaltens stĂŒtzen sich hĂ€ufig auf aggregierte Daten und vernachlĂ€ssigen dadurch oft HeterogenitĂ€t in Form von Unterschieden zwischen Personen oder Gruppen. Die NichtberĂŒcksichtigung vorliegender HeterogenitĂ€t kann zu verzerrten ParameterschĂ€tzungen und zu falsch positiven oder falsch negativen Tests fĂŒhren. HĂ€ufig kann HeterogenitĂ€t mithilfe von Kovariaten erkannt und vorhergesagt werden. Allerdings erweist sich die Identifizierung von PrĂ€diktoren von HeterogenitĂ€t oft als schwierige Aufgabe. Zur Lösung dieses Problems schlage ich zwei neue AnsĂ€tze vor, um individuelle und gruppenspezifische Unterschiede mithilfe von Kovariaten vorherzusagen. Die vorliegende kumulative Dissertation setzt sich aus drei Projekten zusammen. Projekt 1 widmet sich dem Verfahren IPC-Regression (Individual Parameter Contribution), welches die Exploration von ParameterheterogenitĂ€t in Strukturgleichungsmodellen (SEM) mittels Kovariaten erlaubt. Unter anderem evaluiere ich IPC-Regression fĂŒr dynamische Panel-Modelle, schlage eine alternative SchĂ€tzmethode vor und leite IPCs fĂŒr allgemeine Maximum-Likelihood-SchĂ€tzer her. Projekt 2 veranschaulicht, wie IPC-Regression in der Praxis eingesetzt werden kann. Dazu fĂŒhre ich schrittweise in die Implementierung von IPC-Regression im ipcr-Paket fĂŒr die statistische Programmiersprache R ein. Schließlich werden in Projekt 3 SEM-Trees weiterentwickelt. SEM-Trees sind eine modellbasierte rekursive Partitionierungsmethode zur Identifizierung von Kovariaten, die Gruppenunterschiede in SEM-Parametern vorhersagen. Die bisher verwendeten SEM-Trees sind sehr rechenaufwendig. In Projekt 3 kombiniere ich SEM-Trees mit unterschiedlichen Score-basierten Tests. Die daraus resultierenden Score-Guided-SEM-Tees lassen sich deutlich schneller als herkömmlichen SEM-Trees berechnen und zeigen bessere statistische Eigenschaften.Statistical models of human cognition and behavior often rely on aggregated data and may fail to consider heterogeneity, that is, differences across individuals or groups. If overlooked, heterogeneity can bias parameter estimates and may lead to false-positive or false-negative findings. Often, heterogeneity can be detected and predicted with the help of covariates. However, identifying predictors of heterogeneity can be a challenging task. To solve this issue, I propose two novel approaches for detecting and predicting individual and group differences with covariates. This cumulative dissertation is composed of three projects. Project 1 advances the individual parameter contribution (IPC) regression framework, which allows studying heterogeneity in structural equation model (SEM) parameters by means of covariates. I evaluate the use of IPC regression for dynamic panel models, propose an alternative estimation technique, and derive IPCs for general maximum likelihood estimators. Project 2 illustrates how IPC regression can be used in practice. To this end, I provide a step-by-step introduction to the IPC regression implementation in the ipcr package for the R system for statistical computing. Finally, Project 3 progresses the SEM tree framework. SEM trees are a model-based recursive partitioning method for finding covariates that predict group differences in SEM parameters. Unfortunately, the original SEM tree implementation is computationally demanding. As a solution to this problem, I combine SEM trees with a family of score-based tests. The resulting score-guided SEM trees compute quickly, solving the runtime issues of the original SEM trees, and show favorable statistical properties

    Search Trajectory Networks Applied to the Cyclic Bandwidth Sum Problem

    Get PDF
    Search trajectory networks (STNs) were proposed as a tool to analyze the behavior of metaheuristics in relation to their exploration ability and the search space regions they traverse. The technique derives from the study of fitness landscapes using local optima networks (LONs). STNs are related to LONs in that both are built as graphs, modelling the transitions among solutions or group of solutions in the search space. The key difference is that STN nodes can represent solutions or groups of solutions that are not necessarily locally optimal. This work presents an STN-based study for a particular combinatorial optimization problem, the cyclic bandwidth sum minimization. STNs were employed to analyze the two leading algorithms for this problem: a memetic algorithm and a hyperheuristic memetic algorithm. We also propose a novel grouping method for STNs that can be generally applied to both continuous and combinatorial spaces

    Hierarchical cluster guided labeling: efficient label collection for visual classification

    Get PDF
    2015 Summer.Visual classification is a core component in many visually intelligent systems. For example, recognition of objects and terrains provides perception during path planning and navigation tasks performed by autonomous agents. Supervised visual classifiers are typically trained with large sets of images to yield high classification performance. Although the collection of raw training data is easy, the required human effort to assign labels to this data is time consuming. This is particularly problematic in real-world applications with limited labeling time and resources. Techniques have emerged that are designed to help alleviate the labeling workload but suffer from several shortcomings. First, they do not generalize well to domains with limited a priori knowledge. Second, efficiency is achieved at the cost of collecting significant label noise which inhibits classifier learning or requires additional effort to remove. Finally, they introduce high latency between labeling queries, restricting real-world feasibility. This thesis addresses these shortcomings with unsupervised learning that exploits the hierarchical nature of feature patterns and semantic labels in visual data. Our hierarchical cluster guided labeling (HCGL) framework introduces a novel evaluation of hierarchical groupings to identify the most interesting changes in feature patterns. These changes help localize group selection in the hierarchy to discover and label a spectrum of visual semantics found in the data. We show that employing majority group-based labeling after selection allows HCGL to balance efficiency and label accuracy, yielding higher performing classifiers than other techniques with respect to labeling effort. Finally, we demonstrate the real-world feasibility of our labeling framework by quickly training high performing visual classifiers that aid in successful mobile robot path planning and navigation

    Induction and interaction in the evolution of language and conceptual structure

    Get PDF
    Languages evolve in response to various pressures, and this thesis adopts the view that two pressures are especially important. Firstly, the process of learning a language functions as a pressure for greater simplicity due to a domain-general cognitive preference for simple structure. Secondly, the process of using a language in communicative scenarios functions as a pressure for greater informativeness because ultimately languages are only useful to the extent that they allow their users to express – or indeed represent – nuanced meaning distinctions. These two fundamental properties of language – simplicity and informativeness – are often, but not always, in conflict with each other. In general, a simple language cannot be informative and an informative language cannot be simple, resulting in the simplicity–informativeness tradeoff. Typological studies in several domains, including colour, kinship, and spatial relations, have demonstrated that languages find optimal solutions to this tradeoff – optimal solutions to the problem of balancing, on the one hand, the need for simplicity, and on the other, the need for informativeness. More specifically, the thesis explores how inductive reasoning and communicative interaction contribute to simple and informative structure respectively, with a particular emphasis on how a continuous space of meanings, such as the colour spectrum, may be divided into discrete labelled categories. The thesis first describes information-theoretic perspectives on learning and communication and highlights the fact that one of the hallmark feature of conceptual structure – which I term compactness – is not subject to the simplicity–informativeness tradeoff, since it confers advantages on both learning and use. This means it is unclear whether compact structure derives from a learning pressure or from a communicative pressure. To complicate matters further, some researchers view learning as a pressure for simplicity, as outlined above, while others have argued that learning might function as a pressure for informativeness in the sense that learners might have an a-priori expectation that languages ought to be informative. The thesis attempts to resolve this by formalizing these different perspectives in a model of an idealized Bayesian learner, and this model is used to make specific predictions about how these perspectives will play out during individual concept induction and also during the evolution of conceptual structure over time. Experimental testing of these predictions reveals overwhelming support for the simplicity account: Learners have a preference for simplicity, and over generational time, this preference becomes amplified, ultimately resulting in maximally simple, but nevertheless compact, conceptual structure. This emergent compact structure remains limited, however, because it only permits the expression of a small number of meaning distinctions – the emergent systems become degenerate. This issue is addressed in the second part of the thesis, which compares the outcomes of three experiments. The first replicates the finding above – compact categorical structure emerges from learning; the second and third experiments compare artificial and genuine pressures for expressivity, and they show that it is only in the presence of a live communicative task that higher level structure – a kind of statistical compositionality – can emerge. Working together, the low-level compact categorical structure, derived from learning, and the high-level compositional structure, derived from communicative interaction, provide a solution to the simplicity–informativeness tradeoff, expanding on and lending support to various claims in the literature

    Generalizations of the Multicut Problem for Computer Vision

    Get PDF
    Graph decomposition has always been a very important concept in machine learning and computer vision. Many tasks like image and mesh segmentation, community detection in social networks, as well as object tracking and human pose estimation can be formulated as a graph decomposition problem. The multicut problem in particular is a popular model to optimize for a decomposition of a given graph. Its main advantage is that no prior knowledge about the number of components or their sizes is required. However, it has several limitations, which we address in this thesis: Firstly, the multicut problem allows to specify only cost or reward for putting two direct neighbours into distinct components. This limits the expressibility of the cost function. We introduce special edges into the graph that allow to define cost or reward for putting any two vertices into distinct components, while preserving the original set of feasible solutions. We show that this considerably improves the quality of image and mesh segmentations. Second, multicut is notorious to be NP-hard for general graphs, that limits its applications to small super-pixel graphs. We define and implement two primal feasible heuristics to solve the problem. They do not provide any guarantees on the runtime or quality of solutions, but in practice show good convergence behaviour. We perform an extensive comparison on multiple graphs of different sizes and properties. Third, we extend the multicut framework by introducing node labels, so that we can jointly optimize for graph decomposition and nodes classification by means of exactly the same optimization algorithm, thus eliminating the need to hand-tune optimizers for a particular task. To prove its universality we applied it to diverse computer vision tasks, including human pose estimation, multiple object tracking, and instance-aware semantic segmentation. We show that we can improve the results over the prior art using exactly the same data as in the original works. Finally, we use employ multicuts in two applications: 1) a client-server tool for interactive video segmentation: After the pre-processing of the video a user draws strokes on several frames and a time-coherent segmentation of the entire video is performed on-the-fly. 2) we formulate a method for simultaneous segmentation and tracking of living cells in microscopy data. This task is challenging as cells split and our algorithm accounts for this, creating parental hierarchies. We also present results on multiple model fitting. We find models in data heavily corrupted by noise by finding components defining these models using higher order multicuts. We introduce an interesting extension that allows our optimization to pick better hyperparameters for each discovered model. In summary, this thesis extends the multicut problem in different directions, proposes algorithms for optimization, and applies it to novel data and settings.Die Zerlegung von Graphen ist ein sehr wichtiges Konzept im maschinellen Lernen und maschinellen Sehen. Viele Aufgaben wie Bild- und Gittersegmentierung, KommunitĂ€tserkennung in sozialen Netzwerken, sowie Objektverfolgung und SchĂ€tzung von menschlichen Posen können als Graphzerlegungsproblem formuliert werden. Der Mehrfachschnitt-Ansatz ist ein populĂ€res Mittel um ĂŒber die Zerlegungen eines gegebenen Graphen zu optimieren. Sein grĂ¶ĂŸter Vorteil ist, dass kein Vorwissen ĂŒber die Anzahl an Komponenten und deren GrĂ¶ĂŸen benötigt wird. Dennoch hat er mehrere ernsthafte Limitierungen, welche wir in dieser Arbeit behandeln: Erstens erlaubt der klassische Mehrfachschnitt nur die Spezifikation von Kosten oder Belohnungen fĂŒr die Trennung von zwei Nachbarn in verschiedene Komponenten. Dies schrĂ€nkt die AusdrucksfĂ€higkeit der Kostenfunktion ein und fĂŒhrt zu suboptimalen Ergebnissen. Wir fĂŒgen dem Graphen spezielle Kanten hinzu, welche es erlauben, Kosten oder Belohnungen fĂŒr die Trennung von beliebigen Paaren von Knoten in verschiedene Komponenten zu definieren, ohne die Menge an zulĂ€ssigen Lösungen zu verĂ€ndern. Wir zeigen, dass dies die QualitĂ€t von Bild- und Gittersegmentierungen deutlich verbessert. Zweitens ist das Mehrfachschnittproblem berĂŒchtigt dafĂŒr NP-schwer fĂŒr allgemeine Graphen zu sein, was die Anwendungen auf kleine superpixel-basierte Graphen einschrĂ€nkt. Wir definieren und implementieren zwei primal-zulĂ€ssige Heuristiken um das Problem zu lösen. Diese geben keine Garantien bezĂŒglich der Laufzeit oder der QualitĂ€t der Lösungen, zeigen in der Praxis jedoch gutes Konvergenzverhalten. Wir fĂŒhren einen ausfĂŒhrlichen Vergleich auf vielen Graphen verschiedener GrĂ¶ĂŸen und Eigenschaften durch. Drittens erweitern wir den Mehrfachschnitt-Ansatz um Knoten-Kennzeichnungen, sodass wir gemeinsam ĂŒber Zerlegungen und Knoten-Klassifikationen mit dem gleichen Optimierungs-Algorithmus optimieren können. Dadurch wird der Bedarf der Feinabstimmung einzelner aufgabenspezifischer Löser aus dem Weg gerĂ€umt. Um die AllgemeingĂŒltigkeit dieses Ansatzes zu ĂŒberprĂŒfen, haben wir ihn auf verschiedenen Aufgaben des maschinellen Sehens, einschließlich menschliche PosenschĂ€tzung, Mehrobjektverfolgung und instanz-bewusste semantische Segmentierung, angewandt. Wir zeigen, dass wir Resultate von vorherigen Arbeiten mit exakt den gleichen Daten verbessern können. Abschließend benutzen wir Mehrfachschnitte in zwei Anwendungen: 1) Ein Nutzer-Server-Werkzeug fĂŒr interaktive Video Segmentierung: Nach der Vorbearbeitung eines Videos zeichnet der Nutzer Striche auf mehrere Einzelbilder und eine zeit-kohĂ€rente Segmentierung des gesamten Videos wird in Echtzeit berechnet. 2) Wir formulieren eine Methode fĂŒr simultane Segmentierung und Verfolgung von lebenden Zellen in Mikroskopie-Aufnahmen. Diese Aufgabe ist anspruchsvoll, da Zellen sich aufteilen und unser Algorithmus dies in der Erstellung von Eltern-Hierarchien mitberĂŒcksichtigen muss. Wir prĂ€sentieren außerdem Resultate zur Mehrmodellanpassung. Wir berechnen Modelle in stark verrauschten Daten indem wir mithilfe von Mehrfachschnitten höherer Ordnung Komponenten finden, die diesen Modellen entsprechen. Wir fĂŒhren eine interessante Erweiterung ein, die es unserer Optimierung erlaubt, bessere Hyperparameter fĂŒr jedes entdeckte Modell auszuwĂ€hlen. Zusammenfassend erweitert diese Arbeit den Mehrfachschnitt-Ansatz in unterschiedlichen Richtungen, schlĂ€gt Algorithmen zur Inferenz in den resultierenden Modellen vor und wendet ihn auf neuartigen Daten und Umgebungen an

    A survey of Bayesian Network structure learning

    Get PDF
    • 

    corecore