2,057 research outputs found

    CNN-based fast source device identification

    Full text link
    Source identification is an important topic in image forensics, since it allows to trace back the origin of an image. This represents a precious information to claim intellectual property but also to reveal the authors of illicit materials. In this paper we address the problem of device identification based on sensor noise and propose a fast and accurate solution using convolutional neural networks (CNNs). Specifically, we propose a 2-channel-based CNN that learns a way of comparing camera fingerprint and image noise at patch level. The proposed solution turns out to be much faster than the conventional approach and to ensure an increased accuracy. This makes the approach particularly suitable in scenarios where large databases of images are analyzed, like over social networks. In this vein, since images uploaded on social media usually undergo at least two compression stages, we include investigations on double JPEG compressed images, always reporting higher accuracy than standard approaches

    Localization Techniques for Water Pipeline Leakages: A Review

    Get PDF
    Pipeline leakages in water distribution network (WDN) is one of the prominent issues that has gain an interest among researchers in the past few years. Time and accuracy play an important role in leak localization as it has huge impact to the human population and economic point of view. The complexity of WDN has prompt numerous techniques and methods been introduced focusing on the accuracy and efficacy. In general, localization techniques can be divided into two broad categories; external and internal systems. This paper reviews some of the techniques that has been explored and proposed including the limitations of each techniques. Â

    Detection, Localization and Characterization of Gravitational Wave Bursts in a Pulsar Timing Array

    Get PDF
    Efforts to detect gravitational waves by timing an array of pulsars have focused traditionally on stationary gravitational waves: e.g., stochastic or periodic signals. Gravitational wave bursts --- signals whose duration is much shorter than the observation period --- will also arise in the pulsar timing array waveband. Sources that give rise to detectable bursts include the formation or coalescence of supermassive black holes (SMBHs), the periapsis passage of compact objects in highly elliptic or unbound orbits about a SMBH, or cusps on cosmic strings. Here we describe how pulsar timing array data may be analyzed to detect and characterize these bursts. Our analysis addresses, in a mutually consistent manner, a hierarchy of three questions: \emph{i}) What are the odds that a dataset includes the signal from a gravitational wave burst? \emph{ii}) Assuming the presence of a burst, what is the direction to its source? and \emph{iii}) Assuming the burst propagation direction, what is the burst waveform's time dependence in each of its polarization states? Applying our analysis to synthetic data sets we find that we can \emph{detect} gravitational waves even when the radiation is too weak to either localize the source of infer the waveform, and \emph{detect} and \emph{localize} sources even when the radiation amplitude is too weak to permit the waveform to be determined. While the context of our discussion is gravitational wave detection via pulsar timing arrays, the analysis itself is directly applicable to gravitational wave detection using either ground or space-based detector data.Comment: 43 pages, 13 figures, submitted to ApJ

    Robust PCA as Bilinear Decomposition with Outlier-Sparsity Regularization

    Full text link
    Principal component analysis (PCA) is widely used for dimensionality reduction, with well-documented merits in various applications involving high-dimensional data, including computer vision, preference measurement, and bioinformatics. In this context, the fresh look advocated here permeates benefits from variable selection and compressive sampling, to robustify PCA against outliers. A least-trimmed squares estimator of a low-rank bilinear factor analysis model is shown closely related to that obtained from an 0\ell_0-(pseudo)norm-regularized criterion encouraging sparsity in a matrix explicitly modeling the outliers. This connection suggests robust PCA schemes based on convex relaxation, which lead naturally to a family of robust estimators encompassing Huber's optimal M-class as a special case. Outliers are identified by tuning a regularization parameter, which amounts to controlling sparsity of the outlier matrix along the whole robustification path of (group) least-absolute shrinkage and selection operator (Lasso) solutions. Beyond its neat ties to robust statistics, the developed outlier-aware PCA framework is versatile to accommodate novel and scalable algorithms to: i) track the low-rank signal subspace robustly, as new data are acquired in real time; and ii) determine principal components robustly in (possibly) infinite-dimensional feature spaces. Synthetic and real data tests corroborate the effectiveness of the proposed robust PCA schemes, when used to identify aberrant responses in personality assessment surveys, as well as unveil communities in social networks, and intruders from video surveillance data.Comment: 30 pages, submitted to IEEE Transactions on Signal Processin

    A review of model based and data driven methods targeting hardware systems diagnostics

    Get PDF
    System health diagnosis serves as an underpinning enabler for enhanced safety and optimized maintenance tasks in complex assets. In the past four decades, a wide-range of diagnostic methods have been proposed, focusing either on system or component level. Currently, one of the most quickly emerging concepts within the diagnostic community is system level diagnostics. This approach targets in accurately detecting faults and suggesting to the maintainers a component to be replaced in order to restore the system to a healthy state. System level diagnostics is of great value to complex systems whose downtime due to faults is expensive. This paper aims to provide a comprehensive review of the most recent diagnostics approaches applied to hardware systems. The main objective of this paper is to introduce the concept of system level diagnostics and review and evaluate the collated approaches. In order to achieve this, a comprehensive review of the most recent diagnostic methods implemented for hardware systems or components is conducted, highlighting merits and shortfalls

    Machine Learning Applications for Load Predictions in Electrical Energy Network

    Get PDF
    In this work collected operational data of typical urban and rural energy network are analysed for predictions of energy consumption, as well as for selected region of Nordpool electricity markets. The regression techniques are systematically investigated for electrical energy prediction and correlating other impacting parameters. The k-Nearest Neighbour (kNN), Random Forest (RF) and Linear Regression (LR) are analysed and evaluated both by using continuous and vertical time approach. It is observed that for 30 minutes predictions the RF Regression has the best results, shown by a mean absolute percentage error (MAPE) in the range of 1-2 %. kNN show best results for the day-ahead forecasting with MAPE of 2.61 %. The presented vertical time approach outperforms the continuous time approach. To enhance pre-processing stage, refined techniques from the domain of statistics and time series are adopted in the modelling. Reducing the dimensionality through principal components analysis improves the predictive performance of Recurrent Neural Networks (RNN). In the case of Gated Recurrent Units (GRU) networks, the results for all the seasons are improved through principal components analysis (PCA). This work also considers abnormal operation due to various instances (e.g. random effect, intrusion, abnormal operation of smart devices, cyber-threats, etc.). In the results of kNN, iforest and Local Outlier Factor (LOF) on urban area data and from rural region data, it is observed that the anomaly detection for the scenarios are different. For the rural region, most of the anomalies are observed in the latter timeline of the data concentrated in the last year of the collected data. For the urban area data, the anomalies are spread out over the entire timeline. The frequency of detected anomalies where considerably higher for the rural area load demand than for the urban area load demand. Observing from considered case scenarios, the incidents of detected anomalies are more data driven, than exceptions in the algorithms. It is observed that from the domain knowledge of smart energy systems the LOF is able to detect observations that could not have detected by visual inspection alone, in contrast to kNN and iforest. Whereas kNN and iforest excludes an upper and lower bound, the LOF is density based and separates out anomalies amidst in the data. The capability that LOF has to identify anomalies amidst the data together with the deep domain knowledge is an advantage, when detecting anomalies in smart meter data. This work has shown that the instance based models can compete with models of higher complexity, yet some methods in preprocessing (such as circular coding) does not function for an instance based learner such as k-Nearest Neighbor, and hence kNN can not option for this kind of complexity even in the feature engineering of the model. It will be interesting for the future work of electrical load forecasting to develop solution that combines a high complexity in the feature engineering and have the explainability of instance based models.publishedVersio

    Modeling wind speed with a long-term horizon and high-time interval with a hybrid fourier-neural network model

    Get PDF
    The limited availability of local climatological stations and the limitations to predict the wind speed (WS) accurately are significant barriers to the expansion of wind energy (WE) projects worldwide. A methodology to forecast accurately the WS at the local scale can be used to overcome these barriers. This study proposes a methodology to forecast the WS with high-resolution and long-term horizons, which combines a Fourier model and a nonlinear autoregressive network (NAR). Given the nonlinearities of the WS variations, a NAR model is used to forecast the WS based on the variability identified with the Fourier analysis. The NAR modelled successfully 1.7 years of windspeed with 3 hours of the time interval, what may be considered the longest forecasting horizon with high resolution at the moment
    corecore