638 research outputs found
Accelerating finite state machine-based testing using reinforcement learning
Testing is a crucial phase in the development of complex systems, and this has led to interest in automated test generation techniques based on state-based models. Many approaches use models that are types of finite state machine (FSM). Corresponding test generation algorithms typically require that certain test components, such as reset sequences (RSs) and preset distinguishing sequences (PDSs), have been produced for the FSM specification. Unfortunately, the generation of RSs and PDSs is computationally expensive, and this affects the scalability of such FSM-based test generation algorithms. This paper addresses this scalability problem by introducing a reinforcement learning framework: the Q -Graph framework for MBT. We show how this framework can be used in the generation of RSs and PDSs and consider both (potentially partial) timed and untimed models. The proposed approach was evaluated using three types of FSMs: randomly generated FSMs, FSMs from a benchmark, and an FSM of an Engine Status Manager for a printer. In experiments, the proposed approach was much faster and used much less memory than the state-of-the-art methods in computing PDSs and RSs
La traduzione specializzata all’opera per una piccola impresa in espansione: la mia esperienza di internazionalizzazione in cinese di Bioretics© S.r.l.
Global markets are currently immersed in two all-encompassing and unstoppable processes: internationalization and globalization. While the former pushes companies to look beyond the borders of their country of origin to forge relationships with foreign trading partners, the latter fosters the standardization in all countries, by reducing spatiotemporal distances and breaking down geographical, political, economic and socio-cultural barriers. In recent decades, another domain has appeared to propel these unifying drives: Artificial Intelligence, together with its high technologies aiming to implement human cognitive abilities in machinery. The “Language Toolkit – Le lingue straniere al servizio dell’internazionalizzazione dell’impresa” project, promoted by the Department of Interpreting and Translation (Forlì Campus) in collaboration with the Romagna Chamber of Commerce (Forlì-Cesena and Rimini), seeks to help Italian SMEs make their way into the global market. It is precisely within this project that this dissertation has been conceived. Indeed, its purpose is to present the translation and localization project from English into Chinese of a series of texts produced by Bioretics© S.r.l.: an investor deck, the company website and part of the installation and use manual of the Aliquis© framework software, its flagship product. This dissertation is structured as follows: Chapter 1 presents the project and the company in detail; Chapter 2 outlines the internationalization and globalization processes and the Artificial Intelligence market both in Italy and in China; Chapter 3 provides the theoretical foundations for every aspect related to Specialized Translation, including website localization; Chapter 4 describes the resources and tools used to perform the translations; Chapter 5 proposes an analysis of the source texts; Chapter 6 is a commentary on translation strategies and choices
Lattice Boltzmann method for warm fluid simulations of plasma wakefield acceleration
A comprehensive characterization of lattice Boltzmann (LB) schemes to perform
warm fluid numerical simulations of particle wakefield acceleration (PWFA)
processes is discussed in this paper. The LB schemes we develop hinge on the
moment matching procedure, allowing the fluid description of a warm
relativistic plasma wake generated by a driver pulse propagating in a neutral
plasma. We focus on fluid models equations resulting from two popular closure
assumptions of the relativistic kinetic equations, i.e., the local equilibrium
and the warm plasma closure assumptions. The developed LB schemes can thus be
used to disclose insights on the quantitative differences between the two
closure approaches in the dynamics of PWFA processes. Comparisons between the
proposed schemes and available analytical results are extensively addressed.Comment: 8 figure
A systematic literature review on source code similarity measurement and clone detection: techniques, applications, and challenges
Measuring and evaluating source code similarity is a fundamental software
engineering activity that embraces a broad range of applications, including but
not limited to code recommendation, duplicate code, plagiarism, malware, and
smell detection. This paper proposes a systematic literature review and
meta-analysis on code similarity measurement and evaluation techniques to shed
light on the existing approaches and their characteristics in different
applications. We initially found over 10000 articles by querying four digital
libraries and ended up with 136 primary studies in the field. The studies were
classified according to their methodology, programming languages, datasets,
tools, and applications. A deep investigation reveals 80 software tools,
working with eight different techniques on five application domains. Nearly 49%
of the tools work on Java programs and 37% support C and C++, while there is no
support for many programming languages. A noteworthy point was the existence of
12 datasets related to source code similarity measurement and duplicate codes,
of which only eight datasets were publicly accessible. The lack of reliable
datasets, empirical evaluations, hybrid methods, and focuses on multi-paradigm
languages are the main challenges in the field. Emerging applications of code
similarity measurement concentrate on the development phase in addition to the
maintenance.Comment: 49 pages, 10 figures, 6 table
Tools for efficient Deep Learning
In the era of Deep Learning (DL), there is a fast-growing demand for building and deploying Deep Neural Networks (DNNs) on various platforms. This thesis proposes five tools to address the challenges for designing DNNs that are efficient in time, in resources and in power consumption.
We first present Aegis and SPGC to address the challenges in improving the memory efficiency of DL training and inference. Aegis makes mixed precision training (MPT) stabler by layer-wise gradient scaling. Empirical experiments show that Aegis can improve MPT accuracy by at most 4\%. SPGC focuses on structured pruning: replacing standard convolution with group convolution (GConv) to avoid irregular sparsity. SPGC formulates GConv pruning as a channel permutation problem and proposes a novel heuristic polynomial-time algorithm. Common DNNs pruned by SPGC have maximally 1\% higher accuracy than prior work.
This thesis also addresses the challenges lying in the gap between DNN descriptions and executables by Polygeist for software and POLSCA for hardware. Many novel techniques, e.g. statement splitting and memory partitioning, are explored and used to expand polyhedral optimisation. Polygeist can speed up software execution in sequential and parallel by 2.53 and 9.47 times on Polybench/C. POLSCA achieves 1.5 times speedup over hardware designs directly generated from high-level synthesis on Polybench/C.
Moreover, this thesis presents Deacon, a framework that generates FPGA-based DNN accelerators of streaming architectures with advanced pipelining techniques to address the challenges from heterogeneous convolution and residual connections. Deacon provides fine-grained pipelining, graph-level optimisation, and heuristic exploration by graph colouring. Compared with prior designs, Deacon shows resource/power consumption efficiency improvement of 1.2x/3.5x for MobileNets and 1.0x/2.8x for SqueezeNets.
All these tools are open source, some of which have already gained public engagement. We believe they can make efficient deep learning applications easier to build and deploy.Open Acces
Taylor University Catalog 2023-2024
The 2023-2024 academic catalog of Taylor University in Upland, Indiana.https://pillars.taylor.edu/catalogs/1128/thumbnail.jp
Set-based state estimation and fault diagnosis using constrained zonotopes and applications
This doctoral thesis develops new methods for set-based state estimation and
active fault diagnosis (AFD) of (i) nonlinear discrete-time systems, (ii)
discrete-time nonlinear systems whose trajectories satisfy nonlinear equality
constraints (called invariants), (iii) linear descriptor systems, and (iv)
joint state and parameter estimation of nonlinear descriptor systems. Set-based
estimation aims to compute tight enclosures of the possible system states in
each time step subject to unknown-but-bounded uncertainties. To address this
issue, the present doctoral thesis proposes new methods for efficiently
propagating constrained zonotopes (CZs) through nonlinear mappings. Besides,
this thesis improves the standard prediction-update framework for systems with
invariants using new algorithms for refining CZs based on nonlinear
constraints. In addition, this thesis introduces a new approach for set-based
AFD of a class of nonlinear discrete-time systems. An affine parametrization of
the reachable sets is obtained for the design of an optimal input for set-based
AFD. In addition, this thesis presents new methods based on CZs for set-valued
state estimation and AFD of linear descriptor systems. Linear static
constraints on the state variables can be directly incorporated into CZs.
Moreover, this thesis proposes a new representation for unbounded sets based on
zonotopes, which allows to develop methods for state estimation and AFD also of
unstable linear descriptor systems, without the knowledge of an enclosure of
all the trajectories of the system. This thesis also develops a new method for
set-based joint state and parameter estimation of nonlinear descriptor systems
using CZs in a unified framework. Lastly, this manuscript applies the proposed
set-based state estimation and AFD methods using CZs to unmanned aerial
vehicles, water distribution networks, and a lithium-ion cell.Comment: My PhD Thesis from Federal University of Minas Gerais, Brazil. Most
of the research work has already been published in DOIs
10.1109/CDC.2018.8618678, 10.23919/ECC.2018.8550353,
10.1016/j.automatica.2019.108614, 10.1016/j.ifacol.2020.12.2484,
10.1016/j.ifacol.2021.08.308, 10.1016/j.automatica.2021.109638,
10.1109/TCST.2021.3130534, 10.1016/j.automatica.2022.11042
Quantum mean estimation for lattice field theory
We demonstrate the quantum mean estimation algorithm on Euclidean lattice
field theories. This shows a quadratic advantage over Monte Carlo methods which
persists even in presence of a sign problem, and is insensitive to critical
slowing down. The algorithm is used to compute with and without a sign
problem, a toy U(1) gauge theory model, and the Ising model. The effect of
-gate synthesis errors on a future fault-tolerant quantum computer is
investigated.Comment: 14 pages, 18 figure
Discovering structure without labels
The scarcity of labels combined with an abundance of data makes unsupervised learning more attractive than ever. Without annotations, inductive biases must guide the identification of the most salient structure in the data. This thesis contributes to two aspects of unsupervised learning: clustering and dimensionality reduction.
The thesis falls into two parts. In the first part, we introduce Mod Shift, a clustering method for point data that uses a distance-based notion of attraction and repulsion to determine the number of clusters and the assignment of points to clusters. It iteratively moves points towards crisp clusters like Mean Shift but also has close ties to the Multicut problem via its loss function. As a result, it connects signed graph partitioning to clustering in Euclidean space.
The second part treats dimensionality reduction and, in particular, the prominent neighbor embedding methods UMAP and t-SNE. We analyze the details of UMAP's implementation and find its actual loss function. It differs drastically from the one usually stated. This discrepancy allows us to explain some typical artifacts in UMAP plots, such as the dataset size-dependent tendency to produce overly crisp substructures. Contrary to existing belief, we find that UMAP's high-dimensional similarities are not critical to its success.
Based on UMAP's actual loss, we describe its precise connection to the other state-of-the-art visualization method, t-SNE. The key insight is a new, exact relation between the contrastive loss functions negative sampling, employed by UMAP, and noise-contrastive estimation, which has been used to approximate t-SNE. As a result, we explain that UMAP embeddings appear more compact than t-SNE plots due to increased attraction between neighbors. Varying the attraction strength further, we obtain a spectrum of neighbor embedding methods, encompassing both UMAP- and t-SNE-like versions as special cases. Moving from more attraction to more repulsion shifts the focus of the embedding from continuous, global to more discrete and local structure of the data. Finally, we emphasize the link between contrastive neighbor embeddings and self-supervised contrastive learning. We show that different flavors of contrastive losses can work for both of them with few noise samples
- …