78 research outputs found
Maximizing Revenue in the Presence of Intermediaries
We study the mechanism design problem of selling items to unit-demand
buyers with private valuations for the items. A buyer either participates
directly in the auction or is represented by an intermediary, who represents a
subset of buyers. Our goal is to design robust mechanisms that are independent
of the demand structure (i.e. how the buyers are partitioned across
intermediaries), and perform well under a wide variety of possible contracts
between intermediaries and buyers.
We first study the case of identical items where each buyer draws its
private valuation for an item i.i.d. from a known -regular
distribution. We construct a robust mechanism that, independent of the demand
structure and under certain conditions on the contracts between intermediaries
and buyers, obtains a constant factor of the revenue that the mechanism
designer could obtain had she known the buyers' valuations. In other words, our
mechanism's expected revenue achieves a constant factor of the optimal welfare,
regardless of the demand structure. Our mechanism is a simple posted-price
mechanism that sets a take-it-or-leave-it per-item price that depends on
and the total number of buyers, but does not depend on the demand structure or
the downstream contracts.
Next we generalize our result to the case when the items are not identical.
We assume that the item valuations are separable. For this case, we design a
mechanism that obtains at least a constant fraction of the optimal welfare, by
using a menu of posted prices. This mechanism is also independent of the demand
structure, but makes a relatively stronger assumption on the contracts between
intermediaries and buyers, namely that each intermediary prefers outcomes with
a higher sum of utilities of the subset of buyers represented by it
Analyzing the Efficacy of an LLM-Only Approach for Image-based Document Question Answering
Recent document question answering models consist of two key components: the
vision encoder, which captures layout and visual elements in images, and a
Large Language Model (LLM) that helps contextualize questions to the image and
supplements them with external world knowledge to generate accurate answers.
However, the relative contributions of the vision encoder and the language
model in these tasks remain unclear. This is especially interesting given the
effectiveness of instruction-tuned LLMs, which exhibit remarkable adaptability
to new tasks. To this end, we explore the following aspects in this work: (1)
The efficacy of an LLM-only approach on document question answering tasks (2)
strategies for serializing textual information within document images and
feeding it directly to an instruction-tuned LLM, thus bypassing the need for an
explicit vision encoder (3) thorough quantitative analysis on the feasibility
of such an approach. Our comprehensive analysis encompasses six diverse
benchmark datasets, utilizing LLMs of varying scales. Our findings reveal that
a strategy exclusively reliant on the LLM yields results that are on par with
or closely approach state-of-the-art performance across a range of datasets. We
posit that this evaluation framework will serve as a guiding resource for
selecting appropriate datasets for future research endeavors that emphasize the
fundamental importance of layout and image content information
Is it an i or an l: Test-time Adaptation of Text Line Recognition Models
Recognizing text lines from images is a challenging problem, especially for
handwritten documents due to large variations in writing styles. While text
line recognition models are generally trained on large corpora of real and
synthetic data, such models can still make frequent mistakes if the handwriting
is inscrutable or the image acquisition process adds corruptions, such as
noise, blur, compression, etc. Writing style is generally quite consistent for
an individual, which can be leveraged to correct mistakes made by such models.
Motivated by this, we introduce the problem of adapting text line recognition
models during test time. We focus on a challenging and realistic setting where,
given only a single test image consisting of multiple text lines, the task is
to adapt the model such that it performs better on the image, without any
labels. We propose an iterative self-training approach that uses feedback from
the language model to update the optical model, with confident self-labels in
each iteration. The confidence measure is based on an augmentation mechanism
that evaluates the divergence of the prediction of the model in a local region.
We perform rigorous evaluation of our method on several benchmark datasets as
well as their corrupted versions. Experimental results on multiple datasets
spanning multiple scripts show that the proposed adaptation method offers an
absolute improvement of up to 8% in character error rate with just a few
iterations of self-training at test time
A dedicated greedy pursuit algorithm for sparse spectral representation of music sound
A dedicated algorithm for sparse spectral representation of music sound is presented. The goal is to enable the representation of a piece of music signal as a linear superposition of as few spectral components as possible, without affecting the quality of the reproduction. A representation of this nature is said to be sparse. In the present context sparsity is accomplished by greedy selection of the spectral components, from an overcomplete set called a dictionary. The proposed algorithm is tailored to be applied with trigonometric dictionaries. Its distinctive feature being that it avoids the need for the actual construction of the whole dictionary, by implementing the required operations via the fast Fourier transform. The achieved sparsity is theoretically equivalent to that rendered by the orthogonal matching pursuit (OMP) method. The contribution of the proposed dedicated implementation is to extend the applicability of the standard OMP algorithm, by reducing its storage and computational demands. The suitability of the approach for producing sparse spectral representation is illustrated by comparison with the traditional method, in the line of the short time Fourier transform, involving only the corresponding orthonormal trigonometric basis
Derandomization of auctions
We study the role of randomization in seller optimal (i.e., profit maximization) auctions. Bayesian optimal auctions (e.g., Myerson, 1981) assume that the valuations of the agents are random draws from a distribution and prior-free optimal auctions either are randomized (e.g., Goldberg et al., 2006) or assume the valuations are randomized (e.g., Segal, 2003). Is randomization fundamental to profit maximization in auctions? Our main result is a general approach to derandomize single-item multi-unit unit-demand auctions while approximately preserving their performance (i.e., revenue). Our general technique is constructive but not computationally tractable. We complement the general result with the explicit and computationally-simple derandomization of a particular auction. Our results are obtained through analogy to hat puzzles that are interesting in their own right
Epi InfoTM a mHealth tool for primary field data collection in subsample population of Uttarakhand- A cross sectional study.
Background: A deficient data is among the biggest obstacle facing planners and policy makers. Health data collection in the developing world is often hampered by the high costs and inefficiencies of traditional large-scale paper-based surveys. mHealth using Epi-Info is most appropriate tool to create, share, deploy health surveys and for strengthening of health systems. The program runs on free and open software, is easy to use, and can be downloaded to handheld devices to be used by workers in the field.
Objectives: To find out the usefulness and limitations of data collection for mHealth by use of Epi InfoTM software.
Methods: The devices used Epi Info 7.1.5 (Android version), which has been modeled as a database with variables of the traditional form. A cross sectional survey among adolescents regarding their health needs was carried out in a sample of 200 adolescents (purposive sampling) of rural hilly (Jaunpur block of Tehri Garhwal district) and plain (Doiwala block of Dehradun district) areas of Uttarakhand by the use of Android tablets with Epi InfoTM.
Results: It was found that adolescent questionnaire tool developed in Epi InfoTM android tablet application is a powerful tool for data collection having numerable practical advantages like: Interview Time Tracking (ITT) that gives the reality check in field studies along with cases Geographical presentation by GIS mapping. In addition to this complete filling of data in field so no left over or guessing for data entry operator, paperless, bio-friendly. Despite of Tablet cost, it is cost effective as tablet can be repeatedly used for other surveys.
Conclusions: Epi InfoTM is a developing open access software for primary data collection and analyzing data from the field, with advantageous benefits of epidemiological surveys
- …