2,172 research outputs found
(Psycho-)Analysis of Benchmark Experiments
It is common knowledge that certain characteristics of data sets -- such as linear separability or sample size -- determine the performance of learning algorithms. In this paper we propose a formal framework for investigations on this relationship.
The framework combines three, in their respective scientific discipline well-established, methods. Benchmark experiments are the method of choice in machine and statistical learning to compare algorithms with respect to a certain performance measure on particular data sets. To realize the interaction between data sets and algorithms, the data sets are characterized using statistical and information-theoretic measures; a common approach in the field of meta learning to decide which algorithms are suited to particular data sets. Finally, the performance ranking of algorithms on groups of data sets with similar characteristics is determined by means of recursively partitioning Bradley-Terry models, that are commonly used in psychology to study the preferences of human subjects. The result is a tree with splits in data set characteristics which significantly change the performances of the algorithms. The main advantage is the automatic detection of these important characteristics.
The framework is introduced using a simple artificial example. Its real-word usage is demonstrated by means of an application example consisting of thirteen well-known data sets and six common learning algorithms. All resources to replicate the examples are available online
Case study of a student with an emotional behavioral disorder: an increase in reading fluency and its effect on reading comprehension and behavior in the general education classroom.
This study investigated the effect an increase in reading fluency had on reading comprehension and on-task behavior in the general education classroom. The participant was an 18 year-old with a cognitive disability and emotional behavioral disorder. The intervention included repeated reading, vocabulary, partner reading, comprehension questions, and weekly classroom observations. Data collected throughout the course of the intervention were pre- and post-tests, parent and teacher questionnaires, and measurements of reading fluency, comprehension, and on-task behavior in the classroom. Results showed small gains in reading fluency and on-task behavior in the classroom, but no change in reading comprehension. Some limitations that may have affected progress include: the limited number and length of intervention sessions, amount of time devoted to classroom observations, and participant\u27s attendance. Further research is needed on effective literacy interventions to determine a relationship between reading fluency, comprehension, and on-task behavior in the classroom for students with disabilities
Symplectic Cuts and Projection Quantization
The recently proposed projection quantization, which is a method to quantize
particular subspaces of systems with known quantum theory, is shown to yield a
genuine quantization in several cases. This may be inferred from exact results
established within symplectic cutting.Comment: 12 pages, v2: additional examples and a new reference to related wor
Getting It Right: Employment Subsidy or Minimum Wage?
In monopsony models of the labour market either a minimum wage or an employment
subsidy financed by a lump sum tax on profits can achieve the efficient level of employment and output. Incorporating working conditions into a monopsony model where higher wages raise firm labour supply, but less attractive working conditions reduce it, changes these policy implications. Specifically, a minimum wage policy could, in contrast to an employment subsidy, cause working conditions to deteriorate and welfare to fall. Empirical evidence from
the Republic of Trinidad and Tobago shows that a minimum wage may indeed cause working conditions to worsen
Efficiency Wages and Effort: Are Hard Jobs Better?
Efficiency wage theory predicts that the wage per unit of effort will be lower in intensively monitored sectors. This wage differential will increase in effort. Using employer-employee matched data from Ghana we provide evidence supporting this hypothesis
Identifying Patient-Specific Root Causes with the Heteroscedastic Noise Model
Complex diseases are caused by a multitude of factors that may differ between
patients even within the same diagnostic category. A few underlying root causes
may nevertheless initiate the development of disease within each patient. We
therefore focus on identifying patient-specific root causes of disease, which
we equate to the sample-specific predictivity of the exogenous error terms in a
structural equation model. We generalize from the linear setting to the
heteroscedastic noise model where with
non-linear functions and representing the conditional mean
and mean absolute deviation, respectively. This model preserves identifiability
but introduces non-trivial challenges that require a customized algorithm
called Generalized Root Causal Inference (GRCI) to extract the error terms
correctly. GRCI recovers patient-specific root causes more accurately than
existing alternatives
Sample-Specific Root Causal Inference with Latent Variables
Root causal analysis seeks to identify the set of initial perturbations that
induce an unwanted outcome. In prior work, we defined sample-specific root
causes of disease using exogenous error terms that predict a diagnosis in a
structural equation model. We rigorously quantified predictivity using Shapley
values. However, the associated algorithms for inferring root causes assume no
latent confounding. We relax this assumption by permitting confounding among
the predictors. We then introduce a corresponding procedure called Extract
Errors with Latents (EEL) for recovering the error terms up to contamination by
vertices on certain paths under the linear non-Gaussian acyclic model. EEL also
identifies the smallest sets of dependent errors for fast computation of the
Shapley values. The algorithm bypasses the hard problem of estimating the
underlying causal graph in both cases. Experiments highlight the superior
accuracy and robustness of EEL relative to its predecessors
Algebroid Yang-Mills Theories
A framework for constructing new kinds of gauge theories is suggested.
Essentially it consists in replacing Lie algebras by Lie or Courant algebroids.
Besides presenting novel topological theories defined in arbitrary spacetime
dimensions, we show that equipping Lie algebroids E with a fiber metric having
sufficiently many E-Killing vectors leads to an astonishingly mild deformation
of ordinary Yang-Mills theories: Additional fields turn out to carry no
propagating modes. Instead they serve as moduli parameters gluing together in
part different Yang-Mills theories. This leads to a symmetry enhancement at
critical points of these fields, as is also typical for String effective field
theories.Comment: 4 pages; v3: Minor rewording of v1, version to appear in Phys. Rev.
Let
- …