8,902 research outputs found
Advanced flight control system study
A fly by wire flight control system architecture designed for high reliability includes spare sensor and computer elements to permit safe dispatch with failed elements, thereby reducing unscheduled maintenance. A methodology capable of demonstrating that the architecture does achieve the predicted performance characteristics consists of a hierarchy of activities ranging from analytical calculations of system reliability and formal methods of software verification to iron bird testing followed by flight evaluation. Interfacing this architecture to the Lockheed S-3A aircraft for flight test is discussed. This testbed vehicle can be expanded to support flight experiments in advanced aerodynamics, electromechanical actuators, secondary power systems, flight management, new displays, and air traffic control concepts
Recommended from our members
Using formal methods to support testing
Formal methods and testing are two important approaches that assist in the development of high quality software. While traditionally these approaches have been seen as rivals, in recent
years a new consensus has developed in which they are seen as complementary. This article reviews the state of the art regarding ways in which the presence of a formal specification can be used to assist testing
Inductive machine learning of optimal modular structures: Estimating solutions using support vector machines
Structural optimization is usually handled by iterative methods requiring repeated samples of a physics-based model, but this process can be computationally demanding. Given a set of previously optimized structures of the same topology, this paper uses inductive learning to replace this optimization process entirely by deriving a function that directly maps any given load to an optimal geometry. A support vector machine is trained to determine the optimal geometry of individual modules of a space frame structure given a specified load condition. Structures produced by learning are compared against those found by a standard gradient descent optimization, both as individual modules and then as a composite structure. The primary motivation for this is speed, and results show the process is highly efficient for cases in which similar optimizations must be performed repeatedly. The function learned by the algorithm can approximate the result of optimization very closely after sufficient training, and has also been found effective at generalizing the underlying optima to produce structures that perform better than those found by standard iterative methods
Computation in Finitary Stochastic and Quantum Processes
We introduce stochastic and quantum finite-state transducers as
computation-theoretic models of classical stochastic and quantum finitary
processes. Formal process languages, representing the distribution over a
process's behaviors, are recognized and generated by suitable specializations.
We characterize and compare deterministic and nondeterministic versions,
summarizing their relative computational power in a hierarchy of finitary
process languages. Quantum finite-state transducers and generators are a first
step toward a computation-theoretic analysis of individual, repeatedly measured
quantum dynamical systems. They are explored via several physical systems,
including an iterated beam splitter, an atom in a magnetic field, and atoms in
an ion trap--a special case of which implements the Deutsch quantum algorithm.
We show that these systems' behaviors, and so their information processing
capacity, depends sensitively on the measurement protocol.Comment: 25 pages, 16 figures, 1 table; http://cse.ucdavis.edu/~cmg; numerous
corrections and update
Formal Modeling of Connectionism using Concurrency Theory, an Approach Based on Automata and Model Checking
This paper illustrates a framework for applying formal methods techniques, which are symbolic in nature, to specifying and verifying neural networks, which are sub-symbolic in nature. The paper describes a communicating automata [Bowman & Gomez, 2006] model of neural networks. We also implement the model using timed automata [Alur & Dill, 1994] and then undertake a verification of these models using the model checker Uppaal [Pettersson, 2000] in order to evaluate the performance of learning algorithms. This paper also presents discussion of a number of broad issues concerning cognitive neuroscience and the debate as to whether symbolic processing or connectionism is a suitable representation of cognitive systems. Additionally, the issue of integrating symbolic techniques, such as formal methods, with complex neural networks is discussed. We then argue that symbolic verifications may give theoretically well-founded ways to evaluate and justify neural learning systems in the field of both theoretical research and real world applications
Using quantum theory to reduce the complexity of input-output processes
All natural things process and transform information. They receive
environmental information as input, and transform it into appropriate output
responses. Much of science is dedicated to building models of such systems --
algorithmic abstractions of their input-output behavior that allow us to
simulate how such systems can behave in the future, conditioned on what has
transpired in the past. Here, we show that classical models cannot avoid
inefficiency -- storing past information that is unnecessary for correct future
simulation. We construct quantum models that mitigate this waste, whenever it
is physically possible to do so. This suggests that the complexity of general
input-output processes depends fundamentally on what sort of information theory
we use to describe them.Comment: 10 pages, 5 figure
- …