2,272,494 research outputs found
"The Experimental Film Remake and the Digital Archive Effect: A Movie by Jen Proctor and Man with a Movie Camera: The Global Remake"
This essay explores the notion of the experimental film “remake” and the different spectatorial experiences that arise in watching a canonical experimental film and its digital remake. By examining A Movie by Jen Proctor and Man with a Movie Camera: The Global Remake in relation to their originals, this essay reflects on the different effects produced by these films made in the cinematic and digital eras, respectively, as a result of the different archives of documents available for appropriation and recontextualization at each historical moment. Indeed, I suggest that the original films produce a “material archive effect” while the remakes produce a “digital archive effect,” these effects occurring at the levels of both form and content. I argue that these films offer a point of entry for thinking about how cinematic and digital technologies have each differentially shaped human experience of the “real” as its representation is appropriated and recontextualized
Experimental investigation of an interior search method within a simple framework
A steepest gradient method for solving Linear Programming (LP) problems, followed by a procedure for purifying a non-basic solution to an improved extreme point solution have been embedded within an otherwise simplex based optimiser. The algorithm is designed to be hybrid in nature and exploits many aspects of sparse matrix and revised simplex technology. The interior search step terminates at a boundary point which is usually non-basic. This is then followed by a series of minor pivotal steps which lead to a basic feasible solution with a superior objective function value. It is concluded that the procedures discussed in this paper are likely to have three possible applications, which are
(i) improving a non-basic feasible solution to a superior extreme point solution,
(iii) an improved starting point for the revised simplex method, and
(iii) an efficient implementation of the multiple price strategy of the revised simplex method
Inferring the dynamics of underdamped stochastic systems
Many complex systems, ranging from migrating cells to animal groups, exhibit
stochastic dynamics described by the underdamped Langevin equation. Inferring
such an equation of motion from experimental data can provide profound insight
into the physical laws governing the system. Here, we derive a principled
framework to infer the dynamics of underdamped stochastic systems from
realistic experimental trajectories, sampled at discrete times and subject to
measurement errors. This framework yields an operational method, Underdamped
Langevin Inference (ULI), which performs well on experimental trajectories of
single migrating cells and in complex high-dimensional systems, including
flocks with Viscek-like alignment interactions. Our method is robust to
experimental measurement errors, and includes a self-consistent estimate of the
inference error
The role of modeling in troubleshooting: an example from electronics
Troubleshooting systems is integral to experimental physics in both research
and instructional laboratory settings. The recently adopted AAPT Lab Guidelines
identify troubleshooting as an important learning outcome of the undergraduate
laboratory curriculum. We investigate students' model-based reasoning on a
troubleshooting task using data collected in think-aloud interviews during
which pairs of students attempted to diagnose and repair a malfunctioning
circuit. Our analysis scheme is informed by the Experimental Modeling
Framework, which describes physicists' use of mathematical and conceptual
models when reasoning about experimental systems. We show that this framework
is a useful lens through which to characterize the troubleshooting process.Comment: 4 pages, 2 figures; Submitted to 2015 PERC Proceeding
A framework to create customised LHC analyses within CheckMATE
Checkmate is a framework that allows the user to conveniently test simulated
BSM physics events against current LHC data in order to derive exclusion
limits. For this purpose, the data runs through a detector simulation and is
then processed by a user chosen number of experimental analyses. These analyses
are all defined by signal regions that can be compared to the experimental data
with a multitude of statistical tools. Due to the large and continuously
growing number of experimental analyses available, users may quickly find
themselves in the situation that the study they are particularly interested in
has not (yet) been implemented officially into the Checkmate framework.
However, the code includes a rather simple framework to allow users to add new
analyses on their own. This document serves as a guide to this. In addition,
Checkmate serves as a powerful tool for testing and implementing new search
strategies. To aid this process, many tools are included to allow a rapid
prototyping of new analyses.Comment: 32 pages; V3: Fixed typos, additional comments, updated references,
version submitted to CP
A Factor Framework for Experimental Design for Performance Evaluation of Commercial Cloud Services
Given the diversity of commercial Cloud services, performance evaluations of
candidate services would be crucial and beneficial for both service customers
(e.g. cost-benefit analysis) and providers (e.g. direction of service
improvement). Before an evaluation implementation, the selection of suitable
factors (also called parameters or variables) plays a prerequisite role in
designing evaluation experiments. However, there seems a lack of systematic
approaches to factor selection for Cloud services performance evaluation. In
other words, evaluators randomly and intuitively concerned experimental factors
in most of the existing evaluation studies. Based on our previous taxonomy and
modeling work, this paper proposes a factor framework for experimental design
for performance evaluation of commercial Cloud services. This framework
capsules the state-of-the-practice of performance evaluation factors that
people currently take into account in the Cloud Computing domain, and in turn
can help facilitate designing new experiments for evaluating Cloud services.Comment: 8 pages, Proceedings of the 4th International Conference on Cloud
Computing Technology and Science (CloudCom 2012), pp. 169-176, Taipei,
Taiwan, December 03-06, 201
Isoscalar dipole mode in relativistic random phase approximation
The isoscalar giant dipole resonance structure in Pb is calculated in
the framework of a fully consistent relativistic random phase approximation,
based on effective mean-field Lagrangians with nonlinear meson self-interaction
terms. The results are compared with recent experimental data and with
calculations performed in the Hartree-Fock plus RPA framework. Two basic
isoscalar dipole modes are identified from the analysis of the velocity
distributions. The discrepancy between the calculated strength distributions
and current experimental data is discussed, as well as the implications for the
determination of the nuclear matter incompressibility.Comment: 9 pages, Latex, 3. p.s figs, submitted to Phys. Lett.
- …
