4,454 research outputs found
Support Vector Machine classification of strong gravitational lenses
The imminent advent of very large-scale optical sky surveys, such as Euclid
and LSST, makes it important to find efficient ways of discovering rare objects
such as strong gravitational lens systems, where a background object is
multiply gravitationally imaged by a foreground mass. As well as finding the
lens systems, it is important to reject false positives due to intrinsic
structure in galaxies, and much work is in progress with machine learning
algorithms such as neural networks in order to achieve both these aims. We
present and discuss a Support Vector Machine (SVM) algorithm which makes use of
a Gabor filterbank in order to provide learning criteria for separation of
lenses and non-lenses, and demonstrate using blind challenges that under
certain circumstances it is a particularly efficient algorithm for rejecting
false positives. We compare the SVM engine with a large-scale human examination
of 100000 simulated lenses in a challenge dataset, and also apply the SVM
method to survey images from the Kilo-Degree Survey.Comment: Accepted by MNRA
A network approach to topic models
One of the main computational and scientific challenges in the modern age is
to extract useful information from unstructured texts. Topic models are one
popular machine-learning approach which infers the latent topical structure of
a collection of documents. Despite their success --- in particular of its most
widely used variant called Latent Dirichlet Allocation (LDA) --- and numerous
applications in sociology, history, and linguistics, topic models are known to
suffer from severe conceptual and practical problems, e.g. a lack of
justification for the Bayesian priors, discrepancies with statistical
properties of real texts, and the inability to properly choose the number of
topics. Here we obtain a fresh view on the problem of identifying topical
structures by relating it to the problem of finding communities in complex
networks. This is achieved by representing text corpora as bipartite networks
of documents and words. By adapting existing community-detection methods --
using a stochastic block model (SBM) with non-parametric priors -- we obtain a
more versatile and principled framework for topic modeling (e.g., it
automatically detects the number of topics and hierarchically clusters both the
words and documents). The analysis of artificial and real corpora demonstrates
that our SBM approach leads to better topic models than LDA in terms of
statistical model selection. More importantly, our work shows how to formally
relate methods from community detection and topic modeling, opening the
possibility of cross-fertilization between these two fields.Comment: 22 pages, 10 figures, code available at https://topsbm.github.io
Lorenz, G\"{o}del and Penrose: New perspectives on determinism and causality in fundamental physics
Despite being known for his pioneering work on chaotic unpredictability, the
key discovery at the core of meteorologist Ed Lorenz's work is the link between
space-time calculus and state-space fractal geometry. Indeed, properties of
Lorenz's fractal invariant set relate space-time calculus to deep areas of
mathematics such as G\"{o}del's Incompleteness Theorem. These properties,
combined with some recent developments in theoretical and observational
cosmology, motivate what is referred to as the `cosmological invariant set
postulate': that the universe can be considered a deterministic dynamical
system evolving on a causal measure-zero fractal invariant set in its
state space. Symbolic representations of are constructed explicitly based
on permutation representations of quaternions. The resulting `invariant set
theory' provides some new perspectives on determinism and causality in
fundamental physics. For example, whilst the cosmological invariant set appears
to have a rich enough structure to allow a description of quantum probability,
its measure-zero character ensures it is sparse enough to prevent invariant set
theory being constrained by the Bell inequality (consistent with a partial
violation of the so-called measurement independence postulate). The primacy of
geometry as embodied in the proposed theory extends the principles underpinning
general relativity. As a result, the physical basis for contemporary programmes
which apply standard field quantisation to some putative gravitational
lagrangian is questioned. Consistent with Penrose's suggestion of a
deterministic but non-computable theory of fundamental physics, a
`gravitational theory of the quantum' is proposed based on the geometry of
, with potential observational consequences for the dark universe.Comment: This manuscript has been accepted for publication in Contemporary
Physics and is based on the author's 9th Dennis Sciama Lecture, given in
Oxford and Triest
Insights into cosmological structure formation with machine learning
Our modern understanding of cosmological structure formation posits that small matter density fluctuations present in the early Universe, as traced by the cosmic microwave background, grow via gravitational instability to form extended haloes of dark matter. A theoretical understanding of the structure, evolution and formation of dark matter haloes is an essential step towards unravelling the intricate connection between halo and galaxy formation, needed to test our cosmological model against data from upcoming galaxy surveys. Physical understanding of the process of dark matter halo formation is made difficult by the highly non-linear nature of the haloes' evolution. I describe a new approach to gain physical insight into cosmological structure formation based on machine learning. This approach combines the ability of machine learning algorithms to learn non-linear relationships, with techniques that enable us to physically interpret the learnt mapping. I describe applications of the method, with the aim of investigating which aspects of the early universe density field impact the later formation of dark matter haloes. First I present a case where the process of halo formation is turned into a binary classification problem; the algorithm predicts whether or not dark matter `particles' in the initial conditions of a simulation will collapse into haloes of a given mass range. Second, I present its generalization to regression, where the algorithm infers the final mass of the halo to which each particle will later belong. I show that the initial tidal shear does not play a significant role compared to the initial density field in establishing final halo masses. Finally, I demonstrate that extending the framework to deep learning algorithms such as convolutional neural networks allows us to explore connections between the early universe and late time haloes beyond those studied by existing analytic approximations of halo collapse
Quantitative electron microscopy for microstructural characterisation
Development of materials for high-performance applications requires accurate and useful analysis tools. In parallel with advances in electron microscopy hardware, we require analysis approaches to better understand microstructural behaviour. Such improvements in characterisation capability permit informed alloy design.
New approaches to the characterisation of metallic materials are presented, primarily using signals collected from electron microscopy experiments. Electron backscatter diffraction is regularly used to investigate crystallography in the scanning electron microscope, and combined with energy-dispersive X-ray spectroscopy to simultaneusly investigate chemistry. New algorithms and analysis pipelines are developed to permit accurate and routine microstructural evaluation, leveraging a variety of machine learning approaches. This thesis investigates the structure and behaviour of Co/Ni-base superalloys, derived from V208C. Use of the presently developed techniques permits informed development of a new generation of advanced gas turbine engine materials.Open Acces
Artificial Intelligence in Materials Science: Applications of Machine Learning to Extraction of Physically Meaningful Information from Atomic Resolution Microscopy Imaging
Materials science is the cornerstone for technological development of the modern world that has been largely shaped by the advances in fabrication of semiconductor materials and devices. However, the Moore’s Law is expected to stop by 2025 due to reaching the limits of traditional transistor scaling. However, the classical approach has shown to be unable to keep up with the needs of materials manufacturing, requiring more than 20 years to move a material from discovery to market. To adapt materials fabrication to the needs of the 21st century, it is necessary to develop methods for much faster processing of experimental data and connecting the results to theory, with feedback flow in both directions. However, state-of-the-art analysis remains selective and manual, prone to human error and unable to handle large quantities of data generated by modern equipment. Recent advances in scanning transmission electron and scanning tunneling microscopies have allowed imaging and manipulation of materials on the atomic level, and these capabilities require development of automated, robust, reproducible methods.Artificial intelligence and machine learning have dealt with similar issues in applications to image and speech recognition, autonomous vehicles, and other projects that are beginning to change the world around us. However, materials science faces significant challenges preventing direct application of the such models without taking physical constraints and domain expertise into account.Atomic resolution imaging can generate data that can lead to better understanding of materials and their properties through using artificial intelligence methods. Machine learning, in particular combinations of deep learning and probabilistic modeling, can learn to recognize physical features in imaging, making this process automated and speeding up characterization. By incorporating the knowledge from theory and simulations with such frameworks, it is possible to create the foundation for the automated atomic scale manufacturing
- …