3,249 research outputs found
Tracking Vector Magnetograms with the Magnetic Induction Equation
The differential affine velocity estimator (DAVE) developed in Schuck (2006)
for estimating velocities from line-of-sight magnetograms is modified to
directly incorporate horizontal magnetic fields to produce a differential
affine velocity estimator for vector magnetograms (DAVE4VM). The DAVE4VM's
performance is demonstrated on the synthetic data from the anelastic
pseudospectral ANMHD simulations that were used in the recent comparison of
velocity inversion techniques by Welsch (2007). The DAVE4VM predicts roughly
95% of the helicity rate and 75% of the power transmitted through the
simulation slice. Inter-comparison between DAVE4VM and DAVE and further
analysis of the DAVE method demonstrates that line-of-sight tracking methods
capture the shearing motion of magnetic footpoints but are insensitive to flux
emergence -- the velocities determined from line-of-sight methods are more
consistent with horizontal plasma velocities than with flux transport
velocities. These results suggest that previous studies that rely on velocities
determined from line-of-sight methods such as the DAVE or local correlation
tracking may substantially misrepresent the total helicity rates and power
through the photosphere.Comment: 30 pages, 13 figure
Troubles with Bayesianism: An introduction to the psychological immune system
A Bayesian mind is, at its core, a rational mind. Bayesianism is thus well-suited to predict and explain mental processes that best exemplify our ability to be rational. However, evidence from belief acquisition and change appears to show that we do not acquire and update information in a Bayesian way. Instead, the principles of belief acquisition and updating seem grounded in maintaining a psychological immune system rather than in approximating
a Bayesian processor
Direct calculation of the hard-sphere crystal/melt interfacial free energy
We present a direct calculation by molecular-dynamics computer simulation of
the crystal/melt interfacial free energy, , for a system of hard
spheres of diameter . The calculation is performed by thermodynamic
integration along a reversible path defined by cleaving, using specially
constructed movable hard-sphere walls, separate bulk crystal and fluid systems,
which are then merged to form an interface. We find the interfacial free energy
to be slightly anisotropic with = 0.62, 0.64 and
0.58 for the (100), (110) and (111) fcc crystal/fluid
interfaces, respectively. These values are consistent with earlier density
functional calculations and recent experiments measuring the crystal nucleation
rates from colloidal fluids of polystyrene spheres that have been interpreted
[Marr and Gast, Langmuir {\bf 10}, 1348 (1994)] to give an estimate of
for the hard-sphere system of , slightly lower
than the directly determined value reported here.Comment: 4 pages, 4 figures, submitted to Physical Review Letter
Model-based Cognitive Neuroscience: Multifield Mechanistic Integration in Practice
Autonomist accounts of cognitive science suggest that cognitive model building and theory construction (can or should) proceed independently of findings in neuroscience. Common functionalist justifications of autonomy rely on there being relatively few constraints between neural structure and cognitive function (e.g., Weiskopf, 2011). In contrast, an integrative mechanistic perspective stresses the mutual constraining of structure and function (e.g., Piccinini & Craver, 2011; Povich, 2015). In this paper, I show how model-based cognitive neuroscience (MBCN) epitomizes the integrative mechanistic perspective and concentrates the most revolutionary elements of the cognitive neuroscience revolution (Boone & Piccinini, 2016). I also show how the prominent subset account of functional realization supports the integrative mechanistic perspective I take on MBCN and use it to clarify the intralevel and interlevel components of integration
The Interstellar Rubidium Isotope Ratio toward Rho Ophiuchi A
The isotope ratio, 85Rb/87Rb, places constraints on models of the
nucleosynthesis of heavy elements, but there is no precise determination of the
ratio for material beyond the Solar System. We report the first measurement of
the interstellar Rb isotope ratio. Our measurement of the Rb I line at 7800 A
for the diffuse gas toward rho Oph A yields a value of 1.21 +/- 0.30 (1-sigma)
that differs significantly from the meteoritic value of 2.59. The Rb/K
elemental abundance ratio for the cloud also is lower than that seen in
meteorites. Comparison of the 85Rb/K and 87Rb/K ratios with meteoritic values
indicates that the interstellar 85Rb abundance in this direction is lower than
the Solar System abundance. We attribute the lower abundance to a reduced
contribution from the r-process. Interstellar abundances for Kr, Cd, and Sn are
consistent with much less r-process synthesis for the solar neighborhood
compared to the amount inferred for the Solar System.Comment: 12 pages with 2 figures and 1 table; will appear in ApJ Letter
Learning, Social Intelligence and the Turing Test - why an "out-of-the-box" Turing Machine will not pass the Turing Test
The Turing Test (TT) checks for human intelligence, rather than any putative
general intelligence. It involves repeated interaction requiring learning in
the form of adaption to the human conversation partner. It is a macro-level
post-hoc test in contrast to the definition of a Turing Machine (TM), which is
a prior micro-level definition. This raises the question of whether learning is
just another computational process, i.e. can be implemented as a TM. Here we
argue that learning or adaption is fundamentally different from computation,
though it does involve processes that can be seen as computations. To
illustrate this difference we compare (a) designing a TM and (b) learning a TM,
defining them for the purpose of the argument. We show that there is a
well-defined sequence of problems which are not effectively designable but are
learnable, in the form of the bounded halting problem. Some characteristics of
human intelligence are reviewed including it's: interactive nature, learning
abilities, imitative tendencies, linguistic ability and context-dependency. A
story that explains some of these is the Social Intelligence Hypothesis. If
this is broadly correct, this points to the necessity of a considerable period
of acculturation (social learning in context) if an artificial intelligence is
to pass the TT. Whilst it is always possible to 'compile' the results of
learning into a TM, this would not be a designed TM and would not be able to
continually adapt (pass future TTs). We conclude three things, namely that: a
purely "designed" TM will never pass the TT; that there is no such thing as a
general intelligence since it necessary involves learning; and that
learning/adaption and computation should be clearly distinguished.Comment: 10 pages, invited talk at Turing Centenary Conference CiE 2012,
special session on "The Turing Test and Thinking Machines
Origin Gaps and the Eternal Sunshine of the Second-Order Pendulum
The rich experiences of an intentional, goal-oriented life emerge, in an
unpredictable fashion, from the basic laws of physics. Here I argue that this
unpredictability is no mirage: there are true gaps between life and non-life,
mind and mindlessness, and even between functional societies and groups of
Hobbesian individuals. These gaps, I suggest, emerge from the mathematics of
self-reference, and the logical barriers to prediction that self-referring
systems present. Still, a mathematical truth does not imply a physical one: the
universe need not have made self-reference possible. It did, and the question
then is how. In the second half of this essay, I show how a basic move in
physics, known as renormalization, transforms the "forgetful" second-order
equations of fundamental physics into a rich, self-referential world that makes
possible the major transitions we care so much about. While the universe runs
in assembly code, the coarse-grained version runs in LISP, and it is from that
the world of aim and intention grows.Comment: FQXI Prize Essay 2017. 18 pages, including afterword on
Ostrogradsky's Theorem and an exchange with John Bova, Dresden Craig, and
Paul Livingsto
Multiscale Discriminant Saliency for Visual Attention
The bottom-up saliency, an early stage of humans' visual attention, can be
considered as a binary classification problem between center and surround
classes. Discriminant power of features for the classification is measured as
mutual information between features and two classes distribution. The estimated
discrepancy of two feature classes very much depends on considered scale
levels; then, multi-scale structure and discriminant power are integrated by
employing discrete wavelet features and Hidden markov tree (HMT). With wavelet
coefficients and Hidden Markov Tree parameters, quad-tree like label structures
are constructed and utilized in maximum a posterior probability (MAP) of hidden
class variables at corresponding dyadic sub-squares. Then, saliency value for
each dyadic square at each scale level is computed with discriminant power
principle and the MAP. Finally, across multiple scales is integrated the final
saliency map by an information maximization rule. Both standard quantitative
tools such as NSS, LCC, AUC and qualitative assessments are used for evaluating
the proposed multiscale discriminant saliency method (MDIS) against the
well-know information-based saliency method AIM on its Bruce Database wity
eye-tracking data. Simulation results are presented and analyzed to verify the
validity of MDIS as well as point out its disadvantages for further research
direction.Comment: 16 pages, ICCSA 2013 - BIOCA sessio
- âŠ