4,028 research outputs found
A coherent method for the detection and estimation of continuous gravitational wave signals using a pulsar timing array
The use of a high precision pulsar timing array is a promising approach to
detecting gravitational waves in the very low frequency regime ( Hz) that is complementary to the ground-based efforts (e.g., LIGO,
Virgo) at high frequencies ( Hz) and space-based ones (e.g.,
LISA) at low frequencies ( Hz). One of the target sources for
pulsar timing arrays are individual supermassive black hole binaries that are
expected to form in galactic mergers. In this paper, a likelihood based method
for detection and estimation is presented for a monochromatic continuous
gravitational wave signal emitted by such a source. The so-called pulsar terms
in the signal that arise due to the breakdown of the long-wavelength
approximation are explicitly taken into account in this method. In addition,
the method accounts for equality and inequality constraints involved in the
semi-analytical maximization of the likelihood over a subset of the parameters.
The remaining parameters are maximized over numerically using Particle Swarm
Optimization. Thus, the method presented here solves the monochromatic
continuous wave detection and estimation problem without invoking some of the
approximations that have been used in earlier studies.Comment: 33 pages, 10 figures, submitted to Ap
Recommended from our members
Improving the multi-objective evolutionary optimization algorithm for hydropower reservoir operations in the California Oroville-Thermalito complex
This study demonstrates the application of an improved Evolutionary optimization Algorithm (EA), titled Multi-Objective Complex Evolution Global Optimization Method with Principal Component Analysis and Crowding Distance Operator (MOSPD), for the hydropower reservoir operation of the Oroville-Thermalito Complex (OTC) - a crucial head-water resource for the California State Water Project (SWP). In the OTC's water-hydropower joint management study, the nonlinearity of hydropower generation and the reservoir's water elevation-storage relationship are explicitly formulated by polynomial function in order to closely match realistic situations and reduce linearization approximation errors. Comparison among different curve-fitting methods is conducted to understand the impact of the simplification of reservoir topography. In the optimization algorithm development, techniques of crowding distance and principal component analysis are implemented to improve the diversity and convergence of the optimal solutions towards and along the Pareto optimal set in the objective space. A comparative evaluation among the new algorithm MOSPD, the original Multi-Objective Complex Evolution Global Optimization Method (MOCOM), the Multi-Objective Differential Evolution method (MODE), the Multi-Objective Genetic Algorithm (MOGA), the Multi-Objective Simulated Annealing approach (MOSA), and the Multi-Objective Particle Swarm Optimization scheme (MOPSO) is conducted using the benchmark functions. The results show that best the MOSPD algorithm demonstrated the best and most consistent performance when compared with other algorithms on the test problems. The newly developed algorithm (MOSPD) is further applied to the OTC reservoir releasing problem during the snow melting season in 1998 (wet year), 2000 (normal year) and 2001 (dry year), in which the more spreading and converged non-dominated solutions of MOSPD provide decision makers with better operational alternatives for effectively and efficiently managing the OTC reservoirs in response to the different climates, especially drought, which has become more and more severe and frequent in California
Ant colony optimisation-based radiation pattern manipulation algorithm for electronically steerable array radiator antennas
A new algorithm for manipulating the radiation pattern of Electronically Steerable Array Radiator Antennas is proposed. A continuous implementation of the Ant Colony Optimisation (ACO) technique calculates the optimal impedance values of reactances loading different parasitic radiators placed in a circle around a centre antenna. By proposing a method to obtain a suitable sampling frequency of the radiation pattern for use in the optimisation algorithm and by transforming the reactance search space into the search space of associated phases, special care was taken to create a fast and reliable implementation, resulting in an approach that is suitable for real-time implementation. The authors compare their approach to analytical techniques and optimisation algorithms for calculating these reactances. Results show that the method is able to calculate near-optimal solutions for gain optimisation and side lobe reduction
Color Image Segmentation Using the Bee Algorithm in the Markovian Framework
This thesis presents color image segmentation as a vital step of image analysis in computer vision. A survey of the Markov Random Field (MRF) with four different implementation methods for its parameter estimation is provided. In addition, a survey of swarm intelligence and a number of swarm based algorithms are presented. The MRF model is used for color image segmentation in the framework. This thesis introduces a new image segmentation implementation that uses the bee algorithm as an optimization tool in the Markovian framework. The experiments show that the new proposed method performs faster than the existing implementation methods with about the same segmentation accuracy
Classification hardness for supervised learners on 20 years of intrusion detection data
This article consolidates analysis of established (NSL-KDD) and new intrusion detection datasets (ISCXIDS2012, CICIDS2017, CICIDS2018) through the use of supervised machine learning (ML) algorithms. The uniformity in analysis procedure opens up the option to compare the obtained results. It also provides a stronger foundation for the conclusions about the efficacy of supervised learners on the main classification task in network security. This research is motivated in part to address the lack of adoption of these modern datasets. Starting with a broad scope that includes classification by algorithms from different families on both established and new datasets has been done to expand the existing foundation and reveal the most opportune avenues for further inquiry. After obtaining baseline results, the classification task was increased in difficulty, by reducing the available data to learn from, both horizontally and vertically. The data reduction has been included as a stress-test to verify if the very high baseline results hold up under increasingly harsh constraints. Ultimately, this work contains the most comprehensive set of results on the topic of intrusion detection through supervised machine learning. Researchers working on algorithmic improvements can compare their results to this collection, knowing that all results reported here were gathered through a uniform framework. This work's main contributions are the outstanding classification results on the current state of the art datasets for intrusion detection and the conclusion that these methods show remarkable resilience in classification performance even when aggressively reducing the amount of data to learn from
Stream Learning in Energy IoT Systems: A Case Study in Combined Cycle Power Plants
The prediction of electrical power produced in combined cycle power plants is a key challenge in the electrical power and energy systems field. This power production can vary depending on environmental variables, such as temperature, pressure, and humidity. Thus, the business problem is how to predict the power production as a function of these environmental conditions, in order to maximize the profit. The research community has solved this problem by applying Machine Learning techniques, and has managed to reduce the computational and time costs in comparison with the traditional thermodynamical analysis. Until now, this challenge has been tackled from a batch learning perspective, in which data is assumed to be at rest, and where models do not continuously integrate new information into already constructed models. We present an approach closer to the Big Data and Internet of Things paradigms, in which data are continuously arriving and where models learn incrementally, achieving significant enhancements in terms of data processing (time, memory and computational costs), and obtaining competitive performances. This work compares and examines the hourly electrical power prediction of several streaming regressors, and discusses about the best technique in terms of time processing and predictive performance to be applied on this streaming scenario.This work has been partially supported by the EU project iDev40. This project has received funding
from the ECSEL Joint Undertaking (JU) under grant agreement No 783163. The JU receives support from the
European Union’s Horizon 2020 research and innovation programme and Austria, Germany, Belgium, Italy,
Spain, Romania. It has also been supported by the Basque Government (Spain) through the project VIRTUAL
(KK-2018/00096), and by Ministerio de EconomÃa y Competitividad of Spain (Grant Ref. TIN2017-85887-C2-2-P)
- …