78 research outputs found

    Maximizing Revenue in the Presence of Intermediaries

    Get PDF
    We study the mechanism design problem of selling kk items to unit-demand buyers with private valuations for the items. A buyer either participates directly in the auction or is represented by an intermediary, who represents a subset of buyers. Our goal is to design robust mechanisms that are independent of the demand structure (i.e. how the buyers are partitioned across intermediaries), and perform well under a wide variety of possible contracts between intermediaries and buyers. We first study the case of kk identical items where each buyer draws its private valuation for an item i.i.d. from a known λ\lambda-regular distribution. We construct a robust mechanism that, independent of the demand structure and under certain conditions on the contracts between intermediaries and buyers, obtains a constant factor of the revenue that the mechanism designer could obtain had she known the buyers' valuations. In other words, our mechanism's expected revenue achieves a constant factor of the optimal welfare, regardless of the demand structure. Our mechanism is a simple posted-price mechanism that sets a take-it-or-leave-it per-item price that depends on kk and the total number of buyers, but does not depend on the demand structure or the downstream contracts. Next we generalize our result to the case when the items are not identical. We assume that the item valuations are separable. For this case, we design a mechanism that obtains at least a constant fraction of the optimal welfare, by using a menu of posted prices. This mechanism is also independent of the demand structure, but makes a relatively stronger assumption on the contracts between intermediaries and buyers, namely that each intermediary prefers outcomes with a higher sum of utilities of the subset of buyers represented by it

    Analyzing the Efficacy of an LLM-Only Approach for Image-based Document Question Answering

    Full text link
    Recent document question answering models consist of two key components: the vision encoder, which captures layout and visual elements in images, and a Large Language Model (LLM) that helps contextualize questions to the image and supplements them with external world knowledge to generate accurate answers. However, the relative contributions of the vision encoder and the language model in these tasks remain unclear. This is especially interesting given the effectiveness of instruction-tuned LLMs, which exhibit remarkable adaptability to new tasks. To this end, we explore the following aspects in this work: (1) The efficacy of an LLM-only approach on document question answering tasks (2) strategies for serializing textual information within document images and feeding it directly to an instruction-tuned LLM, thus bypassing the need for an explicit vision encoder (3) thorough quantitative analysis on the feasibility of such an approach. Our comprehensive analysis encompasses six diverse benchmark datasets, utilizing LLMs of varying scales. Our findings reveal that a strategy exclusively reliant on the LLM yields results that are on par with or closely approach state-of-the-art performance across a range of datasets. We posit that this evaluation framework will serve as a guiding resource for selecting appropriate datasets for future research endeavors that emphasize the fundamental importance of layout and image content information

    Is it an i or an l: Test-time Adaptation of Text Line Recognition Models

    Full text link
    Recognizing text lines from images is a challenging problem, especially for handwritten documents due to large variations in writing styles. While text line recognition models are generally trained on large corpora of real and synthetic data, such models can still make frequent mistakes if the handwriting is inscrutable or the image acquisition process adds corruptions, such as noise, blur, compression, etc. Writing style is generally quite consistent for an individual, which can be leveraged to correct mistakes made by such models. Motivated by this, we introduce the problem of adapting text line recognition models during test time. We focus on a challenging and realistic setting where, given only a single test image consisting of multiple text lines, the task is to adapt the model such that it performs better on the image, without any labels. We propose an iterative self-training approach that uses feedback from the language model to update the optical model, with confident self-labels in each iteration. The confidence measure is based on an augmentation mechanism that evaluates the divergence of the prediction of the model in a local region. We perform rigorous evaluation of our method on several benchmark datasets as well as their corrupted versions. Experimental results on multiple datasets spanning multiple scripts show that the proposed adaptation method offers an absolute improvement of up to 8% in character error rate with just a few iterations of self-training at test time

    A dedicated greedy pursuit algorithm for sparse spectral representation of music sound

    Get PDF
    A dedicated algorithm for sparse spectral representation of music sound is presented. The goal is to enable the representation of a piece of music signal as a linear superposition of as few spectral components as possible, without affecting the quality of the reproduction. A representation of this nature is said to be sparse. In the present context sparsity is accomplished by greedy selection of the spectral components, from an overcomplete set called a dictionary. The proposed algorithm is tailored to be applied with trigonometric dictionaries. Its distinctive feature being that it avoids the need for the actual construction of the whole dictionary, by implementing the required operations via the fast Fourier transform. The achieved sparsity is theoretically equivalent to that rendered by the orthogonal matching pursuit (OMP) method. The contribution of the proposed dedicated implementation is to extend the applicability of the standard OMP algorithm, by reducing its storage and computational demands. The suitability of the approach for producing sparse spectral representation is illustrated by comparison with the traditional method, in the line of the short time Fourier transform, involving only the corresponding orthonormal trigonometric basis

    Derandomization of auctions

    Get PDF
    We study the role of randomization in seller optimal (i.e., profit maximization) auctions. Bayesian optimal auctions (e.g., Myerson, 1981) assume that the valuations of the agents are random draws from a distribution and prior-free optimal auctions either are randomized (e.g., Goldberg et al., 2006) or assume the valuations are randomized (e.g., Segal, 2003). Is randomization fundamental to profit maximization in auctions? Our main result is a general approach to derandomize single-item multi-unit unit-demand auctions while approximately preserving their performance (i.e., revenue). Our general technique is constructive but not computationally tractable. We complement the general result with the explicit and computationally-simple derandomization of a particular auction. Our results are obtained through analogy to hat puzzles that are interesting in their own right

    Epi InfoTM a mHealth tool for primary field data collection in subsample population of Uttarakhand- A cross sectional study.

    Get PDF
    Background: A deficient data is among the biggest obstacle facing planners and policy makers. Health data collection in the developing world is often hampered by the high costs and inefficiencies of traditional large-scale paper-based surveys. mHealth using Epi-Info is most appropriate tool to create, share, deploy health surveys and for strengthening of health systems. The program runs on free and open software, is easy to use, and can be downloaded to handheld devices to be used by workers in the field.  Objectives: To find out the usefulness and limitations of data collection for mHealth by use of Epi InfoTM software.  Methods: The devices used Epi Info 7.1.5 (Android version), which has been modeled as a database with variables of the traditional form. A cross sectional survey among adolescents regarding their health needs was carried out in a sample of 200 adolescents (purposive sampling) of rural hilly (Jaunpur block of Tehri Garhwal district) and plain (Doiwala block of Dehradun district) areas of Uttarakhand by the use of Android tablets with Epi InfoTM.  Results: It was found that adolescent questionnaire tool developed in Epi InfoTM android tablet application is a powerful tool for data collection having numerable practical advantages like: Interview Time Tracking (ITT) that gives the reality check in field studies along with cases Geographical presentation by GIS mapping. In addition to this complete filling of data in field so no left over or guessing for data entry operator, paperless, bio-friendly. Despite of Tablet cost, it is cost effective as tablet can be repeatedly used for other surveys.  Conclusions: Epi InfoTM is a developing open access software for primary data collection and analyzing data from the field, with advantageous benefits of epidemiological surveys
    • …
    corecore