91 research outputs found

    Using hypergraph theory to model coexistence management and coordinated spectrum allocation for heterogeneous wireless networks operating in shared spectrum

    Get PDF
    Electromagnetic waves in the Radio Frequency (RF) spectrum are used to convey wireless transmissions from one radio antenna to another. Spectrum utilisation factor, which refers to how readily a given spectrum can be reused across space and time while maintaining an acceptable level of transmission errors, is used to measure how efficiently a unit of frequency spectrum can be allocated to a specified number of users. The demand for wireless applications is increasing exponentially, hence there is a need for efficient management of the RF spectrum. However, spectrum usage studies have shown that the spectrum is under-utilised in space and time. A regulatory shift from static spectrum assignment to DSA is one way of addressing this. Licence exemption policy has also been advanced in Dynamic Spectrum Access (DSA) systems to spur wireless innovation and universal access to the internet. Furthermore, there is a shift from homogeneous to heterogeneous radio access and usage of the same spectrum band. These three shifts from traditional spectrum management have led to the challenge of coexistence among heterogeneous wireless networks which access the spectrum using DSA techniques. Cognitive radios have the ability for spectrum agility based on spectrum conditions. However, in the presence of multiple heterogeneous networks and without spectrum coordination, there is a challenge related to switching between available channels to minimise interference and maximise spectrum allocation. This thesis therefore focuses on the design of a framework for coexistence management and spectrum coordination, with the objective of maximising spectrum utilisation across geographical space and across time. The amount of geographical coverage in which a frequency can be used is optimised through frequency reuse while ensuring that harmful interference is minimised. The time during which spectrum is occupied is increased through time-sharing of the same spectrum by two or more networks, while ensuring that spectrum is shared by networks that can coexist in the same spectrum and that the total channel load is not excessive to prevent spectrum starvation. Conventionally, a graph is used to model relationships between entities such as interference relationships among networks. However, the concept of an edge in a graph is not sufficient to model relationships that involve more than two entities, such as more than two networks that are able to share the same channel in the time domain, because an edge can only connect two entities. On the other hand, a hypergraph is a generalisation of an undirected graph in which a hyperedge can connect more than two entities. Therefore, this thesis investigates the use of hypergraph theory to model the RF environment and the spectrum allocation scheme. The hypergraph model was applied to an algorithm for spectrum sharing among 100 heterogeneous wireless networks, whose geo-locations were randomly and independently generated in a 50 km by 50 km area. Simulation results for spectrum utilisation performance have shown that the hypergraph-based model allocated channels, on average, to 8% more networks than the graph-based model. The results also show that, for the same RF environment, the hypergraph model requires up to 36% fewer channels to achieve, on average, 100% operational networks, than the graph model. The rate of growth of the running time of the hypergraph-based algorithm with respect to the input size is equal to the square of the input size, like the graph-based algorithm. Thus, the model achieved better performance at no additional time complexity.Electromagnetic waves in the Radio Frequency (RF) spectrum are used to convey wireless transmissions from one radio antenna to another. Spectrum utilisation factor, which refers to how readily a given spectrum can be reused across space and time while maintaining an acceptable level of transmission errors, is used to measure how efficiently a unit of frequency spectrum can be allocated to a specified number of users. The demand for wireless applications is increasing exponentially, hence there is a need for efficient management of the RF spectrum. However, spectrum usage studies have shown that the spectrum is under-utilised in space and time. A regulatory shift from static spectrum assignment to DSA is one way of addressing this. Licence exemption policy has also been advanced in Dynamic Spectrum Access (DSA) systems to spur wireless innovation and universal access to the internet. Furthermore, there is a shift from homogeneous to heterogeneous radio access and usage of the same spectrum band. These three shifts from traditional spectrum management have led to the challenge of coexistence among heterogeneous wireless networks which access the spectrum using DSA techniques. Cognitive radios have the ability for spectrum agility based on spectrum conditions. However, in the presence of multiple heterogeneous networks and without spectrum coordination, there is a challenge related to switching between available channels to minimise interference and maximise spectrum allocation. This thesis therefore focuses on the design of a framework for coexistence management and spectrum coordination, with the objective of maximising spectrum utilisation across geographical space and across time. The amount of geographical coverage in which a frequency can be used is optimised through frequency reuse while ensuring that harmful interference is minimised. The time during which spectrum is occupied is increased through time-sharing of the same spectrum by two or more networks, while ensuring that spectrum is shared by networks that can coexist in the same spectrum and that the total channel load is not excessive to prevent spectrum starvation. Conventionally, a graph is used to model relationships between entities such as interference relationships among networks. However, the concept of an edge in a graph is not sufficient to model relationships that involve more than two entities, such as more than two networks that are able to share the same channel in the time domain, because an edge can only connect two entities. On the other hand, a hypergraph is a generalisation of an undirected graph in which a hyperedge can connect more than two entities. Therefore, this thesis investigates the use of hypergraph theory to model the RF environment and the spectrum allocation scheme. The hypergraph model was applied to an algorithm for spectrum sharing among 100 heterogeneous wireless networks, whose geo-locations were randomly and independently generated in a 50 km by 50 km area. Simulation results for spectrum utilisation performance have shown that the hypergraph-based model allocated channels, on average, to 8% more networks than the graph-based model. The results also show that, for the same RF environment, the hypergraph model requires up to 36% fewer channels to achieve, on average, 100% operational networks, than the graph model. The rate of growth of the running time of the hypergraph-based algorithm with respect to the input size is equal to the square of the input size, like the graph-based algorithm. Thus, the model achieved better performance at no additional time complexity

    BUILDING EFFICIENT AND COST-EFFECTIVE CLOUD-BASED BIG DATA MANAGEMENT SYSTEMS

    Get PDF
    In today’s big data world, data is being produced in massive volumes, at great velocity and from a variety of different sources such as mobile devices, sensors, a plethora of small devices hooked to the internet (Internet of Things), social networks, communication networks and many others. Interactive querying and large-scale analytics are being increasingly used to derive value out of this big data. A large portion of this data is being stored and processed in the Cloud due the several advantages provided by the Cloud such as scalability, elasticity, availability, low cost of ownership and the overall economies of scale. There is thus, a growing need for large-scale cloud-based data management systems that can support real-time ingest, storage and processing of large volumes of heterogeneous data. However, in the pay-as-you-go Cloud environment, the cost of analytics can grow linearly with the time and resources required. Reducing the cost of data analytics in the Cloud thus remains a primary challenge. In my dissertation research, I have focused on building efficient and cost-effective cloud-based data management systems for different application domains that are predominant in cloud computing environments. In the first part of my dissertation, I address the problem of reducing the cost of transactional workloads on relational databases to support database-as-a-service in the Cloud. The primary challenges in supporting such workloads include choosing how to partition the data across a large number of machines, minimizing the number of distributed transactions, providing high data availability, and tolerating failures gracefully. I have designed, built and evaluated SWORD, an end-to-end scalable online transaction processing system, that utilizes workload-aware data placement and replication to minimize the number of distributed transactions that incorporates a suite of novel techniques to significantly reduce the overheads incurred both during the initial placement of data, and during query execution at runtime. In the second part of my dissertation, I focus on sampling-based progressive analytics as a means to reduce the cost of data analytics in the relational domain. Sampling has been traditionally used by data scientists to get progressive answers to complex analytical tasks over large volumes of data. Typically, this involves manually extracting samples of increasing data size (progressive samples) for exploratory querying. This provides the data scientists with user control, repeatable semantics, and result provenance. However, such solutions result in tedious workflows that preclude the reuse of work across samples. On the other hand, existing approximate query processing systems report early results, but do not offer the above benefits for complex ad-hoc queries. I propose a new progressive data-parallel computation framework, NOW!, that provides support for progressive analytics over big data. In particular, NOW! enables progressive relational (SQL) query support in the Cloud using unique progress semantics that allow efficient and deterministic query processing over samples providing meaningful early results and provenance to data scientists. NOW! enables the provision of early results using significantly fewer resources thereby enabling a substantial reduction in the cost incurred during such analytics. Finally, I propose NSCALE, a system for efficient and cost-effective complex analytics on large-scale graph-structured data in the Cloud. The system is based on the key observation that a wide range of complex analysis tasks over graph data require processing and reasoning about a large number of multi-hop neighborhoods or subgraphs in the graph; examples include ego network analysis, motif counting in biological networks, finding social circles in social networks, personalized recommendations, link prediction, etc. These tasks are not well served by existing vertex-centric graph processing frameworks whose computation and execution models limit the user program to directly access the state of a single vertex, resulting in high execution overheads. Further, the lack of support for extracting the relevant portions of the graph that are of interest to an analysis task and loading it onto distributed memory leads to poor scalability. NSCALE allows users to write programs at the level of neighborhoods or subgraphs rather than at the level of vertices, and to declaratively specify the subgraphs of interest. It enables the efficient distributed execution of these neighborhood-centric complex analysis tasks over largescale graphs, while minimizing resource consumption and communication cost, thereby substantially reducing the overall cost of graph data analytics in the Cloud. The results of our extensive experimental evaluation of these prototypes with several real-world data sets and applications validate the effectiveness of our techniques which provide orders-of-magnitude reductions in the overheads of distributed data querying and analysis in the Cloud

    Mining a Small Medical Data Set by Integrating the Decision Tree and t-test

    Get PDF
    [[abstract]]Although several researchers have used statistical methods to prove that aspiration followed by the injection of 95% ethanol left in situ (retention) is an effective treatment for ovarian endometriomas, very few discuss the different conditions that could generate different recovery rates for the patients. Therefore, this study adopts the statistical method and decision tree techniques together to analyze the postoperative status of ovarian endometriosis patients under different conditions. Since our collected data set is small, containing only 212 records, we use all of these data as the training data. Therefore, instead of using a resultant tree to generate rules directly, we use the value of each node as a cut point to generate all possible rules from the tree first. Then, using t-test, we verify the rules to discover some useful description rules after all possible rules from the tree have been generated. Experimental results show that our approach can find some new interesting knowledge about recurrent ovarian endometriomas under different conditions.[[journaltype]]國外[[incitationindex]]EI[[booktype]]紙本[[countrycodes]]FI

    Interaction Analysis in Smart Work Environments through Fuzzy Temporal Logic

    Get PDF
    Interaction analysis is defined as the generation of situation descriptions from machine perception. World models created through machine perception are used by a reasoning engine based on fuzzy metric temporal logic and situation graph trees, with optional parameter learning and clustering as preprocessing, to deduce knowledge about the observed scene. The system is evaluated in a case study on automatic behavior report generation for staff training purposes in crisis response control rooms

    Visualization of Metabolic Networks

    Get PDF
    The metabolism constitutes the universe of biochemical reactions taking place in a cell of an organism. These processes include the synthesis, transformation, and degradation of molecules for an organism to grow, to reproduce and to interact with its environment. A good way to capture the complexity of these processes is the representation as metabolic network, in which sets of molecules are transformed into products by a chemical reaction, and the products are being processed further. The underlying graph model allows a structural analysis of this network using established graphtheoretical algorithms on the one hand, and a visual representation by applying layout algorithms combined with information visualization techniques on the other. In this thesis we will take a look at three different aspects of graph visualization within the context of biochemical systems: the representation and interactive exploration of static networks, the visual analysis of dynamic networks, and the comparison of two network graphs. We will demonstrate, how established infovis techniques can be combined with new algorithms and applied to specific problems in the area of metabolic network visualization. We reconstruct the metabolic network covering the complete set of chemical reactions present in a generalized eucaryotic cell from real world data available from a popular metabolic pathway data base and present a suitable data structure. As the constructed network is very large, it is not feasible for the display as a whole. Instead, we introduce a technique to analyse this static network in a top-down approach starting with an overview and displaying detailed reaction networks on demand. This exploration method is also applied to compare metabolic networks in different species and from different resources. As for the analysis of dynamic networks, we present a framework to capture changes in the connectivity as well as changes in the attributes associated with the network’s elements

    Interaction Analysis in Smart Work Environments through Fuzzy Temporal Logic

    Get PDF
    Interaction analysis is defined as the generation of situation descriptions from machine perception. World models created through machine perception are used by a reasoning engine based on fuzzy metric temporal logic and situation graph trees, with optional parameter learning and clustering as preprocessing, to deduce knowledge about the observed scene. The system is evaluated in a case study on automatic behavior report generation for staff training purposes in crisis response control rooms

    An Investigation into the Performance Evaluation of Connected Vehicle Applications: From Real-World Experiment to Parallel Simulation Paradigm

    Get PDF
    A novel system was developed that provides drivers lane merge advisories, using vehicle trajectories obtained through Dedicated Short Range Communication (DSRC). It was successfully tested on a freeway using three vehicles, then targeted for further testing, via simulation. The failure of contemporary simulators to effectively model large, complex urban transportation networks then motivated further research into distributed and parallel traffic simulation. An architecture for a closed-loop, parallel simulator was devised, using a new algorithm that accounts for boundary nodes, traffic signals, intersections, road lengths, traffic density, and counts of lanes; it partitions a sample, Tennessee road network more efficiently than tools like METIS, which increase interprocess communications (IPC) overhead by partitioning more transportation corridors. The simulator uses logarithmic accumulation to synchronize parallel simulations, further reducing IPC. Analyses suggest this eliminates up to one-third of IPC overhead incurred by a linear accumulation model

    Optical flow estimation via steered-L1 norm

    Get PDF
    Global variational methods for estimating optical flow are among the best performing methods due to the subpixel accuracy and the ‘fill-in’ effect they provide. The fill-in effect allows optical flow displacements to be estimated even in low and untextured areas of the image. The estimation of such displacements are induced by the smoothness term. The L1 norm provides a robust regularisation term for the optical flow energy function with a very good performance for edge-preserving. However this norm suffers from several issues, among these is the isotropic nature of this norm which reduces the fill-in effect and eventually the accuracy of estimation in areas near motion boundaries. In this paper we propose an enhancement to the L1 norm that improves the fill-in effect for this smoothness term. In order to do this we analyse the structure tensor matrix and use its eigenvectors to steer the smoothness term into components that are ‘orthogonal to’ and ‘aligned with’ image structures. This is done in primal-dual formulation. Results show a reduced end-point error and improved accuracy compared to the conventional L1 norm
    • …
    corecore