17 research outputs found
Competing for Shareable Arms in Multi-Player Multi-Armed Bandits
Competitions for shareable and limited resources have long been studied with
strategic agents. In reality, agents often have to learn and maximize the
rewards of the resources at the same time. To design an individualized
competing policy, we model the competition between agents in a novel
multi-player multi-armed bandit (MPMAB) setting where players are selfish and
aim to maximize their own rewards. In addition, when several players pull the
same arm, we assume that these players averagely share the arms' rewards by
expectation. Under this setting, we first analyze the Nash equilibrium when
arms' rewards are known. Subsequently, we propose a novel SelfishMPMAB with
Averaging Allocation (SMAA) approach based on the equilibrium. We theoretically
demonstrate that SMAA could achieve a good regret guarantee for each player
when all players follow the algorithm. Additionally, we establish that no
single selfish player can significantly increase their rewards through
deviation, nor can they detrimentally affect other players' rewards without
incurring substantial losses for themselves. We finally validate the
effectiveness of the method in extensive synthetic experiments.Comment: ICML 202
Rethinking the Evaluation Protocol of Domain Generalization
Domain generalization aims to solve the challenge of Out-of-Distribution
(OOD) generalization by leveraging common knowledge learned from multiple
training domains to generalize to unseen test domains. To accurately evaluate
the OOD generalization ability, it is necessary to ensure that test data
information is unavailable. However, the current domain generalization protocol
may still have potential test data information leakage. This paper examines the
potential risks of test data information leakage in two aspects of the current
protocol: pretraining on ImageNet and oracle model selection. We propose that
training from scratch and using multiple test domains would result in a more
precise evaluation of OOD generalization ability. We also rerun the algorithms
with the modified protocol and introduce a new leaderboard to encourage future
research in domain generalization with a fairer comparison
Flatness-Aware Minimization for Domain Generalization
Domain generalization (DG) seeks to learn robust models that generalize well
under unknown distribution shifts. As a critical aspect of DG, optimizer
selection has not been explored in depth. Currently, most DG methods follow the
widely used benchmark, DomainBed, and utilize Adam as the default optimizer for
all datasets. However, we reveal that Adam is not necessarily the optimal
choice for the majority of current DG methods and datasets. Based on the
perspective of loss landscape flatness, we propose a novel approach,
Flatness-Aware Minimization for Domain Generalization (FAD), which can
efficiently optimize both zeroth-order and first-order flatness simultaneously
for DG. We provide theoretical analyses of the FAD's out-of-distribution (OOD)
generalization error and convergence. Our experimental results demonstrate the
superiority of FAD on various DG datasets. Additionally, we confirm that FAD is
capable of discovering flatter optima in comparison to other zeroth-order and
first-order flatness-aware optimization methods.Comment: Accepted by ICCV202
On the Out-Of-Distribution Generalization of Multimodal Large Language Models
We investigate the generalization boundaries of current Multimodal Large
Language Models (MLLMs) via comprehensive evaluation under out-of-distribution
scenarios and domain-specific tasks. We evaluate their zero-shot generalization
across synthetic images, real-world distributional shifts, and specialized
datasets like medical and molecular imagery. Empirical results indicate that
MLLMs struggle with generalization beyond common training domains, limiting
their direct application without adaptation. To understand the cause of
unreliable performance, we analyze three hypotheses: semantic
misinterpretation, visual feature extraction insufficiency, and mapping
deficiency. Results identify mapping deficiency as the primary hurdle. To
address this problem, we show that in-context learning (ICL) can significantly
enhance MLLMs' generalization, opening new avenues for overcoming
generalization barriers. We further explore the robustness of ICL under
distribution shifts and show its vulnerability to domain shifts, label shifts,
and spurious correlation shifts between in-context examples and test data
Recommended from our members
Opening the black box: user-log analyses of children’s e-Book reading and associations with word knowledge
Salient syllabi: Examining design characteristics of science online courses in higher education.
The importance of online learning in higher education settings is growing, not only in wake of the Covid-19 pandemic. Therefore, metrics to evaluate and increase the quality of online instruction are crucial for improving student learning. Whereas instructional quality is traditionally evaluated with course observations or student evaluations, course syllabi offer a novel approach to predict course quality even prior to the first day of classes. This study develops an online course design characteristics rubric for science course syllabi. Utilizing content analysis, inductive coding, and deductive coding, we established four broad high-quality course design categories: course organization, course objectives and alignment, interpersonal interactions, and technology. Additionally, this study exploratively applied the rubric on 11 online course syllabi (N = 635 students) and found that these design categories explained variation in student performance
Effects of Biogas Slurry Recirculation on Anaerobic Digestion Performance of Maize Straw Silage
In order to investigate the effects of slurry recirculation technology on anaerobic digestion performance of maize straw silage, maize straw silage was fermented with recirculated biogas slurry, and the gas production, pH value, methane content, volatile organic acids (VFAs) contents, chemical oxygen demand (COD) removal rate and other indicators were studied. The results showed that the fermentation time was positively correlated with daily gas production, methane content, cumulative gas production, VFAs and COD removal rate. Although the pH value fluctuated, it was still in the normal reaction range. The daily gas production was about 1.26 L. The acetic acid content increased first, then decreased, then increased, and finally stabilized. The biogas slurry recirculation technology saves water resources by 40 mL/d without affecting the normal gas production of anaerobic fermentation, and reduces the consumption of environmental resources. It has important development significance for the sustainable use of biomass resources
Stable Learning via Sparse Variable Independence
The problem of covariate-shift generalization has attracted intensive research attention. Previous stable learning algorithms employ sample reweighting schemes to decorrelate the covariates when there is no explicit domain information about training data. However, with finite samples, it is difficult to achieve the desirable weights that ensure perfect independence to get rid of the unstable variables. Besides, decorrelating within stable variables may bring about high variance of learned models because of the over-reduced effective sample size. A tremendous sample size is required for these algorithms to work. In this paper, with theoretical justification, we propose SVI (Sparse Variable Independence) for the covariate-shift generalization problem. We introduce sparsity constraint to compensate for the imperfectness of sample reweighting under the finite-sample setting in previous methods. Furthermore, we organically combine independence-based sample reweighting and sparsity-based variable selection in an iterative way to avoid decorrelating within stable variables, increasing the effective sample size to alleviate variance inflation. Experiments on both synthetic and real-world datasets demonstrate the improvement of covariate-shift generalization performance brought by SVI
Covariate-Shift Generalization via Random Sample Weighting
Shifts in the marginal distribution of covariates from training to the test phase, named covariate-shifts, often lead to unstable prediction performance across agnostic testing data, especially under model misspecification. Recent literature on invariant learning attempts to learn an invariant predictor from heterogeneous environments. However, the performance of the learned predictor depends heavily on the availability and quality of provided environments. In this paper, we propose a simple and effective non-parametric method for generating heterogeneous environments via Random Sample Weighting (RSW). Given the training dataset from a single source environment, we randomly generate a set of covariate-determining sample weights and use each weighted training distribution to simulate an environment. We theoretically show that under appropriate conditions, such random sample weighting can produce sufficient heterogeneity to be exploited by common invariance constraints to find the invariant variables for stable prediction under covariate shifts. Extensive experiments on both simulated and real-world datasets clearly validate the effectiveness of our method