448,935 research outputs found
Cellular Automata Applications in Shortest Path Problem
Cellular Automata (CAs) are computational models that can capture the
essential features of systems in which global behavior emerges from the
collective effect of simple components, which interact locally. During the last
decades, CAs have been extensively used for mimicking several natural processes
and systems to find fine solutions in many complex hard to solve computer
science and engineering problems. Among them, the shortest path problem is one
of the most pronounced and highly studied problems that scientists have been
trying to tackle by using a plethora of methodologies and even unconventional
approaches. The proposed solutions are mainly justified by their ability to
provide a correct solution in a better time complexity than the renowned
Dijkstra's algorithm. Although there is a wide variety regarding the
algorithmic complexity of the algorithms suggested, spanning from simplistic
graph traversal algorithms to complex nature inspired and bio-mimicking
algorithms, in this chapter we focus on the successful application of CAs to
shortest path problem as found in various diverse disciplines like computer
science, swarm robotics, computer networks, decision science and biomimicking
of biological organisms' behaviour. In particular, an introduction on the first
CA-based algorithm tackling the shortest path problem is provided in detail.
After the short presentation of shortest path algorithms arriving from the
relaxization of the CAs principles, the application of the CA-based shortest
path definition on the coordinated motion of swarm robotics is also introduced.
Moreover, the CA based application of shortest path finding in computer
networks is presented in brief. Finally, a CA that models exactly the behavior
of a biological organism, namely the Physarum's behavior, finding the
minimum-length path between two points in a labyrinth is given.Comment: To appear in the book: Adamatzky, A (Ed.) Shortest path solvers. From
software to wetware. Springer, 201
Spectra of "Real-World" Graphs: Beyond the Semi-Circle Law
Many natural and social systems develop complex networks, that are usually
modelled as random graphs. The eigenvalue spectrum of these graphs provides
information about their structural properties. While the semi-circle law is
known to describe the spectral density of uncorrelated random graphs, much less
is known about the eigenvalues of real-world graphs, describing such complex
systems as the Internet, metabolic pathways, networks of power stations,
scientific collaborations or movie actors, which are inherently correlated and
usually very sparse. An important limitation in addressing the spectra of these
systems is that the numerical determination of the spectra for systems with
more than a few thousand nodes is prohibitively time and memory consuming.
Making use of recent advances in algorithms for spectral characterization, here
we develop new methods to determine the eigenvalues of networks comparable in
size to real systems, obtaining several surprising results on the spectra of
adjacency matrices corresponding to models of real-world graphs. We find that
when the number of links grows as the number of nodes, the spectral density of
uncorrelated random graphs does not converge to the semi-circle law.
Furthermore, the spectral densities of real-world graphs have specific features
depending on the details of the corresponding models. In particular, scale-free
graphs develop a triangle-like spectral density with a power law tail, while
small-world graphs have a complex spectral density function consisting of
several sharp peaks. These and further results indicate that the spectra of
correlated graphs represent a practical tool for graph classification and can
provide useful insight into the relevant structural properties of real
networks.Comment: 14 pages, 9 figures (corrected typos, added references) accepted for
Phys. Rev.
Developing Tools for Networks of Processors
A great deal of research eort is currently being made in the realm of so called natural computing. Natural computing mainly focuses on the denition, formal description, analysis, simulation and programming of new models of computation (usually with the same expressive power as Turing Machines) inspired by Nature, which makes them particularly suitable for the simulation of complex systems.Some of the best known natural computers are Lindenmayer systems (Lsystems, a kind of grammar with parallel derivation), cellular automata, DNA computing, genetic and evolutionary algorithms, multi agent systems, arti- cial neural networks, P-systems (computation inspired by membranes) and NEPs (or networks of evolutionary processors). This chapter is devoted to this last model
Optimization Approaches To Protect Transportation Infrastructure Against Strategic and Random Disruptions
Past and recent events have proved that critical infrastructure are vulnerable to natural catastrophes, unintentional accidents and terrorist attacks. Protecting these systems is critical to avoid loss of life and to guard against economical upheaval. A systematic approach to plan security investments is paramount to guarantee that limited protection resources are utilized in the most effcient manner. This thesis provides a detailed review of the optimization models that have been introduced in the past to identify vulnerabilities and protection plans for critical infrastructure. The main objective of this thesis is to study new and more realistic models to protect transportation infrastructure such as railway and road systems against man made and natural disruptions. Solution algorithms are devised to effciently solve the complex formulations proposed. Finally, several illustrative case studies are analysed to demonstrate how solving these models can be used to support effcient protection decisions
Understanding cities with machine eyes: A review of deep computer vision in urban analytics
Modelling urban systems has interested planners and modellers for decades. Different models have been achieved relying on mathematics, cellular automation, complexity, and scaling. While most of these models tend to be a simplification of reality, today within the paradigm shifts of artificial intelligence across the different fields of science, the applications of computer vision show promising potential in understanding the realistic dynamics of cities. While cities are complex by nature, computer vision shows progress in tackling a variety of complex physical and non-physical visual tasks. In this article, we review the tasks and algorithms of computer vision and their applications in understanding cities. We attempt to subdivide computer vision algorithms into tasks, and cities into layers to show evidence of where computer vision is intensively applied and where further research is needed. We focus on highlighting the potential role of computer vision in understanding urban systems related to the built environment, natural environment, human interaction, transportation, and infrastructure. After showing the diversity of computer vision algorithms and applications, the challenges that remain in understanding the integration between these different layers of cities and their interactions with one another relying on deep learning and computer vision. We also show recommendations for practice and policy-making towards reaching AI-generated urban policies
Visual Question Answering
There has been immense progress in the fields of computer vision, object detection and natural language processing (NLP) in recent years. Artificial Intelligence (AI) systems, such as question answering models, use NLP to provide a comprehensive capability to the machine. Such a machine can answer natural language queries about any portion of an unstructured text. An extension of this system is to combine NLP with computer vision to accomplish the task of Visual Question Answering (VQA), which is to build a system that can answer natural language queries about images. A number of systems have been proposed for VQA that use deep-learning architectures and learning algorithms. This project introduces a VQA system that gains deep understanding from images by using a deep convolutional neural network (CNN) that extracts image features. More specifically, the feature embeddings from the output layer of the VGG19 [1] model are used for this purpose. Our system achieves complex reasoning capabilities and understanding of natural language, so that it can correctly interpret the query and return an appropriate answer. The InferSent [2] model is used for obtaining sentence level embeddings to extract features from the question. Different architectures are proposed to merge the image and language models. Our system achieves results comparable to the baseline systems on the VQA dataset
- …