18,849 research outputs found
A fuzzy set theory-based fast fault diagnosis approach for rotators of induction motors
Induction motors have been widely used in industry, agriculture, transportation, national defense engineering, etc. Defects of the motors will not only cause the abnormal operation of production equipment but also cause the motor to run in a state of low energy efficiency before evolving into a fault shutdown. The former may lead to the suspension of the production process, while the latter may lead to additional energy loss. This paper studies a fuzzy rule-based expert system for this purpose and focuses on the analysis of many knowledge representation methods and reasoning techniques. The rotator fault of induction motors is analyzed and diagnosed by using this knowledge, and the diagnosis result is displayed. The simulation model can effectively simulate the broken rotator fault by changing the resistance value of the equivalent rotor winding. And the influence of the broken rotor bar fault on the motors is described, which provides a basis for the fault characteristics analysis. The simulation results show that the proposed method can realize fast fault diagnosis for rotators of induction motors
Green Carbon Footprint for Model Inference Serving via Exploiting Mixed-Quality Models and GPU Partitioning
This paper presents a solution to the challenge of mitigating carbon
emissions from large-scale high performance computing (HPC) systems and
datacenters that host machine learning (ML) inference services. ML inference is
critical to modern technology products, but it is also a significant
contributor to datacenter compute cycles and carbon emissions. We introduce
Clover, a carbon-friendly ML inference service runtime system that balances
performance, accuracy, and carbon emissions through mixed-quality models and
GPU resource partitioning. Our experimental results demonstrate that Clover is
effective in substantially reducing carbon emissions while maintaining high
accuracy and meeting service level agreement (SLA) targets. Therefore, it is a
promising solution toward achieving carbon neutrality in HPC systems and
datacenters
Identifying Appropriate Intellectual Property Protection Mechanisms for Machine Learning Models: A Systematization of Watermarking, Fingerprinting, Model Access, and Attacks
The commercial use of Machine Learning (ML) is spreading; at the same time,
ML models are becoming more complex and more expensive to train, which makes
Intellectual Property Protection (IPP) of trained models a pressing issue.
Unlike other domains that can build on a solid understanding of the threats,
attacks and defenses available to protect their IP, the ML-related research in
this regard is still very fragmented. This is also due to a missing unified
view as well as a common taxonomy of these aspects.
In this paper, we systematize our findings on IPP in ML, while focusing on
threats and attacks identified and defenses proposed at the time of writing. We
develop a comprehensive threat model for IP in ML, categorizing attacks and
defenses within a unified and consolidated taxonomy, thus bridging research
from both the ML and security communities
Many Physical Design Problems are Sparse QCQPs
Physical design refers to mathematical optimization of a desired objective
(e.g. strong light--matter interactions, or complete quantum state transfer)
subject to the governing dynamical equations, such as Maxwell's or
Schrodinger's differential equations. Computing an optimal design is
challenging: generically, these problems are highly nonconvex and finding
global optima is NP hard. Here we show that for linear-differential-equation
dynamics (as in linear electromagnetism, elasticity, quantum mechanics, etc.),
the physical-design optimization problem can be transformed to a sparse-matrix,
quadratically constrained quadratic program (QCQP). Sparse QCQPs can be tackled
with convex optimization techniques (such as semidefinite programming) that
have thrived for identifying global bounds and high-performance designs in
other areas of science and engineering, but seemed inapplicable to the design
problems of wave physics. We apply our formulation to prototypical photonic
design problems, showing the possibility to compute fundamental limits for
large-area metasurfaces, as well as the identification of designs approaching
global optimality. Looking forward, our approach highlights the promise of
developing bespoke algorithms tailored to specific physical design problems.Comment: 9 pages, 4 figures, plus references and Supplementary Material
Quantifying the Expressive Capacity of Quantum Systems: Fundamental Limits and Eigentasks
The expressive capacity of quantum systems for machine learning is limited by
quantum sampling noise incurred during measurement. Although it is generally
believed that noise limits the resolvable capacity of quantum systems, the
precise impact of noise on learning is not yet fully understood. We present a
mathematical framework for evaluating the available expressive capacity of
general quantum systems from a finite number of measurements, and provide a
methodology for extracting the extrema of this capacity, its eigentasks.
Eigentasks are a native set of functions that a given quantum system can
approximate with minimal error. We show that extracting low-noise eigentasks
leads to improved performance for machine learning tasks such as
classification, displaying robustness to overfitting. We obtain a tight bound
on the expressive capacity, and present analyses suggesting that correlations
in the measured quantum system enhance learning capacity by reducing noise in
eigentasks. These results are supported by experiments on superconducting
quantum processors. Our findings have broad implications for quantum machine
learning and sensing applications.Comment: 7 + 21 pages, 4 + 12 figures, 1 tabl
Colour technologies for content production and distribution of broadcast content
The requirement of colour reproduction has long been a priority driving the development of new colour imaging systems that maximise human perceptual plausibility. This thesis explores machine learning algorithms for colour processing to assist both content production and distribution. First, this research studies colourisation technologies with practical use cases in restoration and processing of archived content. The research targets practical deployable solutions, developing a cost-effective pipeline which integrates the activity of the producer into the processing workflow. In particular, a fully automatic image colourisation paradigm using Conditional GANs is proposed to improve content generalisation and colourfulness of existing baselines. Moreover, a more conservative solution is considered by providing references to guide the system towards more accurate colour predictions. A fast-end-to-end architecture is proposed to improve existing exemplar-based image colourisation methods while decreasing the complexity and runtime. Finally, the proposed image-based methods are integrated into a video colourisation pipeline. A general framework is proposed to reduce the generation of temporal flickering or propagation of errors when such methods are applied frame-to-frame. The proposed model is jointly trained to stabilise the input video and to cluster their frames with the aim of learning scene-specific modes. Second, this research explored colour processing technologies for content distribution with the aim to effectively deliver the processed content to the broad audience. In particular, video compression is tackled by introducing a novel methodology for chroma intra prediction based on attention models. Although the proposed architecture helped to gain control over the reference samples and better understand the prediction process, the complexity of the underlying neural network significantly increased the encoding and decoding time. Therefore, aiming at efficient deployment within the latest video coding standards, this work also focused on the simplification of the proposed architecture to obtain a more compact and explainable model
Endogenous measures for contextualising large-scale social phenomena: a corpus-based method for mediated public discourse
This work presents an interdisciplinary methodology for developing endogenous measures of group membership through analysis of pervasive linguistic patterns in public discourse. Focusing on political discourse, this work critiques the conventional approach to the study of political participation, which is premised on decontextualised, exogenous measures to characterise groups. Considering the theoretical and empirical weaknesses of decontextualised approaches to large-scale social phenomena, this work suggests that contextualisation using endogenous measures might provide a complementary perspective to mitigate such weaknesses.
This work develops a sociomaterial perspective on political participation in mediated discourse as affiliatory action performed through language. While the affiliatory function of language is often performed consciously (such as statements of identity), this work is concerned with unconscious features (such as patterns in lexis and grammar). This work argues that pervasive patterns in such features that emerge through socialisation are resistant to change and manipulation, and thus might serve as endogenous measures of sociopolitical contexts, and thus of groups.
In terms of method, the work takes a corpus-based approach to the analysis of data from the Twitter messaging service whereby patterns in users’ speech are examined statistically in order to trace potential community membership. The method is applied in the US state of Michigan during the second half of 2018—6 November having been the date of midterm (i.e. non-Presidential) elections in the United States. The corpus is assembled from the original posts of 5,889 users, who are nominally geolocalised to 417 municipalities. These users are clustered according to pervasive language features. Comparing the linguistic clusters according to the municipalities they represent finds that there are regular sociodemographic differentials across clusters. This is understood as an indication of social structure, suggesting that endogenous measures derived from pervasive patterns in language may indeed offer a complementary, contextualised perspective on large-scale social phenomena
Efficient inference in the transverse field Ising model
In this paper we introduce an approximate method to solve the quantum cavity
equations for transverse field Ising models. The method relies on a projective
approximation of the exact cavity distributions of imaginary time trajectories
(paths). A key feature, novel in the context of similar algorithms, is the
explicit separation of the classical and quantum parts of the distributions.
Numerical simulations show accurate results in comparison with the sampled
solution of the cavity equations, the exact diagonalization of the Hamiltonian
(when possible) and other approximate inference methods in the literature. The
computational complexity of this new algorithm scales linearly with the
connectivity of the underlying lattice, enabling the study of highly connected
networks, as the ones often encountered in quantum machine learning problems
Qluster: An easy-to-implement generic workflow for robust clustering of health data
The exploration of heath data by clustering algorithms allows to better describe the populations of interest by seeking the sub-profiles that compose it. This therefore reinforces medical knowledge, whether it is about a disease or a targeted population in real life. Nevertheless, contrary to the so-called conventional biostatistical methods where numerous guidelines exist, the standardization of data science approaches in clinical research remains a little discussed subject. This results in a significant variability in the execution of data science projects, whether in terms of algorithms used, reliability and credibility of the designed approach. Taking the path of parsimonious and judicious choice of both algorithms and implementations at each stage, this article proposes Qluster, a practical workflow for performing clustering tasks. Indeed, this workflow makes a compromise between (1) genericity of applications (e.g. usable on small or big data, on continuous, categorical or mixed variables, on database of high-dimensionality or not), (2) ease of implementation (need for few packages, few algorithms, few parameters, ...), and (3) robustness (e.g. use of proven algorithms and robust packages, evaluation of the stability of clusters, management of noise and multicollinearity). This workflow can be easily automated and/or routinely applied on a wide range of clustering projects. It can be useful both for data scientists with little experience in the field to make data clustering easier and more robust, and for more experienced data scientists who are looking for a straightforward and reliable solution to routinely perform preliminary data mining. A synthesis of the literature on data clustering as well as the scientific rationale supporting the proposed workflow is also provided. Finally, a detailed application of the workflow on a concrete use case is provided, along with a practical discussion for data scientists. An implementation on the Dataiku platform is available upon request to the authors
- …