2,057 research outputs found
CNN-based fast source device identification
Source identification is an important topic in image forensics, since it
allows to trace back the origin of an image. This represents a precious
information to claim intellectual property but also to reveal the authors of
illicit materials. In this paper we address the problem of device
identification based on sensor noise and propose a fast and accurate solution
using convolutional neural networks (CNNs). Specifically, we propose a
2-channel-based CNN that learns a way of comparing camera fingerprint and image
noise at patch level. The proposed solution turns out to be much faster than
the conventional approach and to ensure an increased accuracy. This makes the
approach particularly suitable in scenarios where large databases of images are
analyzed, like over social networks. In this vein, since images uploaded on
social media usually undergo at least two compression stages, we include
investigations on double JPEG compressed images, always reporting higher
accuracy than standard approaches
Localization Techniques for Water Pipeline Leakages: A Review
Pipeline leakages in water distribution network (WDN) is one of the prominent issues that has gain an interest among researchers in the past few years. Time and accuracy play an important role in leak localization as it has huge impact to the human population and economic point of view. The complexity of WDN has prompt numerous techniques and methods been introduced focusing on the accuracy and efficacy. In general, localization techniques can be divided into two broad categories; external and internal systems. This paper reviews some of the techniques that has been explored and proposed including the limitations of each techniques. Â
Detection, Localization and Characterization of Gravitational Wave Bursts in a Pulsar Timing Array
Efforts to detect gravitational waves by timing an array of pulsars have
focused traditionally on stationary gravitational waves: e.g., stochastic or
periodic signals. Gravitational wave bursts --- signals whose duration is much
shorter than the observation period --- will also arise in the pulsar timing
array waveband. Sources that give rise to detectable bursts include the
formation or coalescence of supermassive black holes (SMBHs), the periapsis
passage of compact objects in highly elliptic or unbound orbits about a SMBH,
or cusps on cosmic strings. Here we describe how pulsar timing array data may
be analyzed to detect and characterize these bursts. Our analysis addresses, in
a mutually consistent manner, a hierarchy of three questions: \emph{i}) What
are the odds that a dataset includes the signal from a gravitational wave
burst? \emph{ii}) Assuming the presence of a burst, what is the direction to
its source? and \emph{iii}) Assuming the burst propagation direction, what is
the burst waveform's time dependence in each of its polarization states?
Applying our analysis to synthetic data sets we find that we can \emph{detect}
gravitational waves even when the radiation is too weak to either localize the
source of infer the waveform, and \emph{detect} and \emph{localize} sources
even when the radiation amplitude is too weak to permit the waveform to be
determined. While the context of our discussion is gravitational wave detection
via pulsar timing arrays, the analysis itself is directly applicable to
gravitational wave detection using either ground or space-based detector data.Comment: 43 pages, 13 figures, submitted to ApJ
Robust PCA as Bilinear Decomposition with Outlier-Sparsity Regularization
Principal component analysis (PCA) is widely used for dimensionality
reduction, with well-documented merits in various applications involving
high-dimensional data, including computer vision, preference measurement, and
bioinformatics. In this context, the fresh look advocated here permeates
benefits from variable selection and compressive sampling, to robustify PCA
against outliers. A least-trimmed squares estimator of a low-rank bilinear
factor analysis model is shown closely related to that obtained from an
-(pseudo)norm-regularized criterion encouraging sparsity in a matrix
explicitly modeling the outliers. This connection suggests robust PCA schemes
based on convex relaxation, which lead naturally to a family of robust
estimators encompassing Huber's optimal M-class as a special case. Outliers are
identified by tuning a regularization parameter, which amounts to controlling
sparsity of the outlier matrix along the whole robustification path of (group)
least-absolute shrinkage and selection operator (Lasso) solutions. Beyond its
neat ties to robust statistics, the developed outlier-aware PCA framework is
versatile to accommodate novel and scalable algorithms to: i) track the
low-rank signal subspace robustly, as new data are acquired in real time; and
ii) determine principal components robustly in (possibly) infinite-dimensional
feature spaces. Synthetic and real data tests corroborate the effectiveness of
the proposed robust PCA schemes, when used to identify aberrant responses in
personality assessment surveys, as well as unveil communities in social
networks, and intruders from video surveillance data.Comment: 30 pages, submitted to IEEE Transactions on Signal Processin
A review of model based and data driven methods targeting hardware systems diagnostics
System health diagnosis serves as an underpinning enabler for enhanced safety and optimized maintenance tasks in complex assets. In the past four decades, a wide-range of diagnostic methods have been proposed, focusing either on system or component level. Currently, one of the most quickly emerging concepts within the diagnostic community is system level diagnostics. This approach targets in accurately detecting faults and suggesting to the maintainers a component to be replaced in order to restore the system to a healthy state. System level diagnostics is of great value to complex systems whose downtime due to faults is expensive. This paper aims to provide a comprehensive review of the most recent diagnostics approaches applied to hardware systems. The main objective of this paper is to introduce the concept of system level diagnostics and review and evaluate the collated approaches. In order to achieve this, a comprehensive review of the most recent diagnostic methods implemented for hardware systems or components is conducted, highlighting merits and shortfalls
Machine Learning Applications for Load Predictions in Electrical Energy Network
In this work collected operational data of typical urban and rural energy network are analysed for predictions of energy consumption, as well as for selected region of Nordpool electricity markets. The regression techniques are systematically investigated for electrical energy prediction and correlating other impacting parameters. The k-Nearest Neighbour (kNN), Random Forest (RF) and Linear Regression (LR) are analysed and evaluated both by using continuous and vertical time approach. It is observed that for 30 minutes predictions the RF Regression has the best results, shown by a mean absolute percentage error (MAPE) in the range of 1-2 %. kNN show best results for the day-ahead forecasting with MAPE of 2.61 %. The presented vertical time approach outperforms the continuous time approach. To enhance pre-processing stage, refined techniques from the domain of statistics and time series are adopted in the modelling. Reducing the dimensionality through principal components analysis improves the predictive performance of Recurrent Neural Networks (RNN). In the case of Gated Recurrent Units (GRU) networks, the results for all the seasons are improved through principal components analysis (PCA). This work also considers abnormal operation due to various instances (e.g. random effect, intrusion, abnormal operation of smart devices, cyber-threats, etc.). In the results of kNN, iforest and Local Outlier Factor (LOF) on urban area data and from rural region data, it is observed that the anomaly detection for the scenarios are different. For the rural region, most of the anomalies are observed in the latter timeline of the data concentrated in the last year of the collected data. For the urban area data, the anomalies are spread out over the entire timeline. The frequency of detected anomalies where considerably higher for the rural area load demand than for the urban area load demand. Observing from considered case scenarios, the incidents of detected anomalies are more data driven, than exceptions in the algorithms. It is observed that from the domain knowledge of smart energy systems the LOF is able to detect observations that could not have detected by visual inspection alone, in contrast to kNN and iforest. Whereas kNN and iforest excludes an upper and lower bound, the LOF is density based and separates out anomalies amidst in the data. The capability that LOF has to identify anomalies amidst the data together with the deep domain knowledge is an advantage, when detecting anomalies in smart meter data. This work has shown that the instance based models can compete with models of higher complexity, yet some methods in preprocessing (such as circular coding) does not function for an instance based learner such as k-Nearest Neighbor, and hence kNN can not option for this kind of complexity even in the feature engineering of the model. It will be interesting for the future work of electrical load forecasting to develop solution that combines a high complexity in the feature engineering and have the explainability of instance based models.publishedVersio
Modeling wind speed with a long-term horizon and high-time interval with a hybrid fourier-neural network model
The limited availability of local climatological stations and the limitations to predict the wind speed (WS) accurately are significant barriers to the expansion of wind energy (WE) projects worldwide. A methodology to forecast accurately the WS at the local scale can be used to overcome these barriers. This study proposes a methodology to forecast the WS with high-resolution and long-term horizons, which combines a Fourier model and a nonlinear autoregressive network (NAR). Given the nonlinearities of the WS variations, a NAR model is used to forecast the WS based on the variability identified with the Fourier analysis. The NAR modelled successfully 1.7 years of windspeed with 3 hours of the time interval, what may be considered the longest forecasting horizon with high resolution at the moment
- …