1,001 research outputs found
Computational Complexity for Physicists
These lecture notes are an informal introduction to the theory of
computational complexity and its links to quantum computing and statistical
mechanics.Comment: references updated, reprint available from
http://itp.nat.uni-magdeburg.de/~mertens/papers/complexity.shtm
Scalable fault management architecture for dynamic optical networks : an information-theoretic approach
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2008.MIT Barker Engineering Library copy: printed in pages.Also issued printed in pages.Includes bibliographical references (leaves 255-262).All-optical switching, in place of electronic switching, of high data-rate lightpaths at intermediate nodes is one of the key enabling technologies for economically scalable future data networks. This replacement of electronic switching with optical switching at intermediate nodes, however, presents new challenges for fault detection and localization in reconfigurable all-optical networks. Presently, fault detection and localization techniques, as implemented in SONET/G.709 networks, rely on electronic processing of parity checks at intermediate nodes. If similar techniques are adapted to all-optical reconfigurable networks, optical signals need to be tapped out at intermediate nodes for parity checks. This additional electronic processing would break the all-optical transparency paradigm and thus significantly diminish the cost advantages of all-optical networks. In this thesis, we propose new fault-diagnosis approaches specifically tailored to all-optical networks, with an objective of keeping the diagnostic capital expenditure and the diagnostic operation effort low. Instead of the aforementioned passive monitoring paradigm based on parity checks, we propose a proactive lightpath probing paradigm: optical probing signals are sent along a set of lightpaths in the network, and network state (i.e., failure pattern) is then inferred from testing results of this set of end-to-end lightpath measurements. Moreover, we assume that a subset of network nodes (up to all the nodes) is equipped with diagnostic agents - including both transmitters/receivers for probe transmission/detection and software processes for probe management to perform fault detection and localization. The design objectives of this proposed proactive probing paradigm are two folded: i) to minimize the number of lightpath probes to keep the diagnostic operational effort low, and ii) to minimize the number of diagnostic hardware to keep the diagnostic capital expenditure low.(cont.) The network fault-diagnosis problem can be mathematically modeled with a group testing-over-graphs framework. In particular, the network is abstracted as a graph in which the failure status of each node/link is modeled with a random variable (e.g. Bernoulli distribution). A probe over any path in the graph results in a value, defined as the probe syndrome, which is a function of all the random variables associated in that path. A network failure pattern is inferred through a set of probe syndromes resulting from a set of optimally chosen probes. This framework enriches the traditional group-testing problem by introducing a topological structure, and can be extended to model many other network-monitoring problems (e.g., packet delay, packet drop ratio, noise and etc) by choosing appropriate state variables. Under the group-testing-over-graphs framework with a probabilistic failure model, we initiate an information-theoretic approach to minimizing the average number of lightpath probes to identify all possible network failure patterns. Specifically, we have established an isomorphic mapping between the fault-diagnosis problem in network management and the source-coding problem in Information Theory. This mapping suggests that the minimum average number of lightpath probes required is lower bounded by the information entropy of the network state and efficient source-coding algorithms (e.g. the run-length code) can be translated into scalable fault-diagnosis schemes under some additional probe feasibility constraint. Our analytical and numerical investigations yield a guideline for designing scalable fault-diagnosis algorithms: each probe should provide approximately 1-bit of state information, and thus the total number of probes required is approximately equal to the entropy of the network state.(cont.) To address the hardware cost of diagnosis, we also developed a probabilistic analysis framework to characterize the trade-off between hardware cost (i.e., the number of nodes equipped with Tx/Rx pairs) and diagnosis capability (i.e., the probability of successful failure detection and localization). Our results suggest that, for practical situations, the hardware cost can be reduced significantly by accepting a small amount of uncertainty about the failure status.by Yonggang Wen.Ph.D
General equilibrium characteristics of a dual-lift helicopter system
The equilibrium characteristics of a dual-lift helicopter system are examined. The system consists of the cargo attached by cables to the endpoints of a spreader bar which is suspended by cables below two helicopters. Results are given for the orientation angles of the suspension system and its internal forces, and for the helicopter thrust vector requirements under general circumstances, including nonidentical helicopters, any accelerating or static equilibrium reference flight condition, any system heading relative to the flight direction, and any distribution of the load to the two helicopters. Optimum tether angles which minimize the sum of the required thrust magnitudes are also determined. The analysis does not consider the attitude degrees of freedom of the load and helicopters in detail, but assumes that these bodies are stable, and that their aerodynamic forces in equilibrium flight can be determined independently as functions of the reference trajectory. The ranges of these forces for sample helicopters and loads are examined and their effects on the equilibrium characteristics are given parametrically in the results
Recommended from our members
Improving fault coverage and minimising the cost of fault identification when testing from finite state machines
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Software needs to be adequately tested in order to increase the confidence that the system being developed is reliable. However, testing is a complicated and expensive process. Formal specification based models such as finite state machines have been widely used in system modelling and testing. In this PhD thesis, we primarily investigate fault detection and identification when testing from finite state machines. The research in this thesis is mainly comprised of three topics - construction of multiple Unique Input/Output (UIO) sequences using Metaheuristic Optimisation Techniques (MOTs), the improved fault
coverage by using robust Unique Input/Output Circuit (UIOC) sequences, and fault diagnosis when testing from finite state machines. In the studies of the construction of UIOs, a model is proposed where a fitness function is defined to guide the search for input sequences that are potentially UIOs. In the studies of the improved fault coverage, a new type of UIOCs is defined. Based upon the Rural Chinese Postman Algorithm (RCPA), a new approach is proposed for the construction of more robust test sequences. In the studies of fault diagnosis, heuristics are defined that attempt to lead to failures being observed in some shorter test sequences, which helps to reduce the
cost of fault isolation and identification. The proposed approaches and techniques were evaluated with regard to a set of case studies, which provides experimental evidence for their efficacy.Brunel Research Initiative and Enterprise Fund (BRIEF) Award from Brunel University and Departmental bursary from Department of Information Systems and Computing, Brunel Universit
A Distribution Network Reconfiguration and Islanding Strategy
With the development of Smart Grid, the reliability and stability of the power system are significantly improved. However, a large-scale outage still possibly occurs when the power system is exposed to extreme conditions. Power system blackstart, the restoration after a complete or partial outage is a key issue needed to be studied for the safety of power system. Network reconfiguration is one of the most important steps when crews try to rapidly restore the network. Therefore, planning an optimal network reconfiguration scheme with the most efficient restoration target at the primary stage of system restoration is necessary and it also builds the foundation to the following restoration process. Besides, the utilization of distributed generators (DGs) has risen sharply in the power system and it plays a critical role in the future Smart Grid to modernize the power grid. The emerging Smart Grid technology, which enables self-sufficient power systems with DGs, provides further opportunities to enhance self-healing capability. The introduction of DGs makes a quick and efficient restoration of power system possible.
In this thesis, based on the topological characteristics of scale-free networks and the Discrete Particle Swarm Optimization (DPSO) algorithm, a network reconfiguration scheme is proposed. A power system structure can be converted into a system consisting of nodes and edges. Indices that reflect the nodes’ and edges’ topological characteristics in Graph Theory can be utilized to describe the importance of loads and transmission lines in the power system. Therefore, indices like node importance degree, line betweenness centrality and clustering coefficient are introduced to weigh the importance of loads and transmission lines. Based on these indices, an objective function which aims to restore as many important loads and transmission lines as possible and also subjected to constraints is formulated. The effectiveness of potential reconfiguration scheme is verified by Depth First Search (DFS) algorithm. Finally, DPSO algorithm is employed to obtain the optimal reconfiguration scheme. The comprehensive reconfiguration scheme proposed by my thesis can be the theoretical basis for the power grid dispatchers.
Besides, DGs are introduced in this thesis to enhance the restoration efficiency and success rate at the primary stage of network restoration. Firstly, the selection and classification principle of DGs are introduced in my thesis. In addition, the start sequence principle of DGs is presented as a foundation for the following stability analysis of network restoration with DGs. Then, the objective function subjected to constraints that aims to restore as many important loads as possible is formulated. Based on the restoration objective, islands that include part of important and restorable loads are formed because the DGs’ capacity cannot ensure an entire restoration of the outage areas. Finally, DPSO is used to obtain the optimal solution of islanding strategy and the state sequence matrix is utilized to represent the solution space.
It is believed that this work will provide some useful insight into improving the power system resiliency in the face of extreme events such as natural or man-made disasters
Geometric data understanding : deriving case specific features
There exists a tradition using precise geometric modeling, where uncertainties in data can be considered noise. Another tradition relies on statistical nature of vast quantity of data, where geometric regularity is intrinsic to data and statistical models usually grasp this level only indirectly. This work focuses on point cloud data of natural resources and the silhouette recognition from video input as two real world examples of problems having geometric content which is intangible at the raw data presentation.
This content could be discovered and modeled to some degree by such machine learning (ML) approaches like deep learning, but either a direct coverage of geometry in samples or addition of special geometry invariant layer is necessary. Geometric content is central when there is a need for direct observations of spatial variables, or one needs to gain a mapping to a geometrically consistent data representation, where e.g. outliers or noise can be easily discerned.
In this thesis we consider transformation of original input data to a geometric feature space in two example problems. The first example is curvature of surfaces, which has met renewed interest since the introduction of ubiquitous point cloud data and the maturation of the discrete differential geometry. Curvature spectra can characterize a spatial sample rather well, and provide useful features for ML purposes. The second example involves projective methods used to video stereo-signal analysis in swimming analytics.
The aim is to find meaningful local geometric representations for feature generation, which also facilitate additional analysis based on geometric understanding of the model. The features are associated directly to some geometric quantity, and this makes it easier to express the geometric constraints in a natural way, as shown in the thesis. Also, the visualization and further feature generation is much easier. Third, the approach provides sound baseline methods to more traditional ML approaches, e.g. neural network methods. Fourth, most of the ML methods can utilize the geometric features presented in this work as additional features.Geometriassa käytetään perinteisesti tarkkoja malleja, jolloin datassa esiintyvät epätarkkuudet edustavat melua. Toisessa perinteessä nojataan suuren datamäärän tilastolliseen luonteeseen, jolloin geometrinen säännönmukaisuus on datan sisäsyntyinen ominaisuus, joka hahmotetaan tilastollisilla malleilla ainoastaan epäsuorasti. Tämä työ keskittyy kahteen esimerkkiin: luonnonvaroja kuvaaviin pistepilviin ja videohahmontunnistukseen. Nämä ovat todellisia ongelmia, joissa geometrinen sisältö on tavoittamattomissa raakadatan tasolla.
Tämä sisältö voitaisiin jossain määrin löytää ja mallintaa koneoppimisen keinoin, esim. syväoppimisen avulla, mutta joko geometria pitää kattaa suoraan näytteistämällä tai tarvitaan neuronien lisäkerros geometrisia invariansseja varten. Geometrinen sisältö on keskeinen, kun tarvitaan suoraa avaruudellisten suureiden havainnointia, tai kun tarvitaan kuvaus geometrisesti yhtenäiseen dataesitykseen, jossa poikkeavat näytteet tai melu voidaan helposti erottaa.
Tässä työssä tarkastellaan datan muuntamista geometriseen piirreavaruuteen kahden esimerkkiohjelman suhteen. Ensimmäinen esimerkki on pintakaarevuus, joka on uudelleen virinneen kiinnostuksen kohde kaikkialle saatavissa olevan datan ja diskreetin geometrian kypsymisen takia. Kaarevuusspektrit voivat luonnehtia avaruudellista kohdetta melko hyvin ja tarjota koneoppimisessa hyödyllisiä piirteitä. Toinen esimerkki koskee projektiivisia menetelmiä käytettäessä stereovideosignaalia uinnin analytiikkaan.
Tavoite on löytää merkityksellisiä paikallisen geometrian esityksiä, jotka samalla mahdollistavat muun geometrian ymmärrykseen perustuvan analyysin. Piirteet liittyvät suoraan johonkin geometriseen suureeseen, ja tämä helpottaa luonnollisella tavalla geometristen rajoitteiden käsittelyä, kuten väitöstyössä osoitetaan. Myös visualisointi ja lisäpiirteiden luonti muuttuu helpommaksi. Kolmanneksi, lähestymistapa suo selkeän vertailumenetelmän perinteisemmille koneoppimisen lähestymistavoille, esim. hermoverkkomenetelmille. Neljänneksi, useimmat koneoppimismenetelmät voivat hyödyntää tässä työssä esitettyjä geometrisia piirteitä lisäämällä ne muiden piirteiden joukkoon
Development of an MSC language and compiler, volume 1
Higher order programming language and compiler for advanced computer software system to be used with manned space flights between 1972 and 198
- …