464 research outputs found
Multi-Prover Commitments Against Non-Signaling Attacks
We reconsider the concept of multi-prover commitments, as introduced in the
late eighties in the seminal work by Ben-Or et al. As was recently shown by
Cr\'{e}peau et al., the security of known two-prover commitment schemes not
only relies on the explicit assumption that the provers cannot communicate, but
also depends on their information processing capabilities. For instance, there
exist schemes that are secure against classical provers but insecure if the
provers have quantum information processing capabilities, and there are schemes
that resist such quantum attacks but become insecure when considering general
so-called non-signaling provers, which are restricted solely by the requirement
that no communication takes place.
This poses the natural question whether there exists a two-prover commitment
scheme that is secure under the sole assumption that no communication takes
place; no such scheme is known.
In this work, we give strong evidence for a negative answer: we show that any
single-round two-prover commitment scheme can be broken by a non-signaling
attack. Our negative result is as bad as it can get: for any candidate scheme
that is (almost) perfectly hiding, there exists a strategy that allows the
dishonest provers to open a commitment to an arbitrary bit (almost) as
successfully as the honest provers can open an honestly prepared commitment,
i.e., with probability (almost) 1 in case of a perfectly sound scheme. In the
case of multi-round schemes, our impossibility result is restricted to
perfectly hiding schemes.
On the positive side, we show that the impossibility result can be
circumvented by considering three provers instead: there exists a three-prover
commitment scheme that is secure against arbitrary non-signaling attacks
Effect of Tuned Parameters on a LSA MCQ Answering Model
This paper presents the current state of a work in progress, whose objective
is to better understand the effects of factors that significantly influence the
performance of Latent Semantic Analysis (LSA). A difficult task, which consists
in answering (French) biology Multiple Choice Questions, is used to test the
semantic properties of the truncated singular space and to study the relative
influence of main parameters. A dedicated software has been designed to fine
tune the LSA semantic space for the Multiple Choice Questions task. With
optimal parameters, the performances of our simple model are quite surprisingly
equal or superior to those of 7th and 8th grades students. This indicates that
semantic spaces were quite good despite their low dimensions and the small
sizes of training data sets. Besides, we present an original entropy global
weighting of answers' terms of each question of the Multiple Choice Questions
which was necessary to achieve the model's success.Comment: 9 page
A Method to Improve the Early Stages of the Robotic Process Automation Lifecycle
The robotic automation of processes is of much interest to
organizations. A common use case is to automate the repetitive manual
tasks (or processes) that are currently done by back-office staff
through some information system (IS). The lifecycle of any Robotic Process
Automation (RPA) project starts with the analysis of the process
to automate. This is a very time-consuming phase, which in practical
settings often relies on the study of process documentation. Such documentation
is typically incomplete or inaccurate, e.g., some documented
cases never occur, occurring cases are not documented, or documented
cases differ from reality. To deploy robots in a production environment
that are designed on such a shaky basis entails a high risk. This paper
describes and evaluates a new proposal for the early stages of an RPA
project: the analysis of a process and its subsequent design. The idea is to
leverage the knowledge of back-office staff, which starts by monitoring
them in a non-invasive manner. This is done through a screen-mousekey-
logger, i.e., a sequence of images, mouse actions, and key actions
are stored along with their timestamps. The log which is obtained in
this way is transformed into a UI log through image-analysis techniques
(e.g., fingerprinting or OCR) and then transformed into a process model
by the use of process discovery algorithms. We evaluated this method for
two real-life, industrial cases. The evaluation shows clear and substantial
benefits in terms of accuracy and speed. This paper presents the method,
along with a number of limitations that need to be addressed such that
it can be applied in wider contexts.Ministerio de EconomÃa y Competitividad TIN2016-76956-C3-2-
Towards an OpenSource Logger for the Analysis of RPA Projects
Process automation typically begins with the observation of
humans conducting the tasks that will be eventually automated. Sim ilarly, successful RPA projects require a prior analysis of the undergo ing processes which are being executed by humans. The process of col lecting this type of information is known as user interface (UI) logging
since it records the interaction against a UI. Main RPA platforms (e.g.,
Blueprism and UIPath) incorporate functionalities that allow the record ing of these UI interactions. However, the records that these platforms
generate lack some functionalities that large-scale RPA projects require.
Besides, they are only understandable by the proper RPA platforms.
This paper presents an extensible and multi-platform OpenSource UI
logger that generate UI logs in a standard format. This system collects
information from all the computers it is running on and sends it to a
central server for its processing. Treatment of the collected information
will allow the creation of an enriched UI log which can be used, among
others purposes, for smart process analysis, machine learning training,
the creation of RPA robots, or, being more general, for task mining .Ministerio de EconomÃa y Competitividad TIN2016-76956-C3-2-R (POLOLAS)Junta de AndalucÃa CEI-12-TIC021Centro para el Desarrollo Tecnol´ogico Industrial (CDTI) P011-19/E0
A non-intrusive movie recommendation system
Several recommendation systems have been developed to support the user in choosing an interesting movie from multimedia repositories. The widely utilized collaborative-filtering systems focus on the analysis of user profiles or user ratings of the items. However, these systems decrease their performance at the start-up phase and due to privacy issues, when a user hides most of his personal data. On the other hand, content-based recommendation systems compare movie features to suggest similar multimedia contents; these systems are based on less invasive observations, however they find some difficulties to supply tailored suggestions. In this paper, we propose a plot-based recommendation system, which is based upon an evaluation of similarity among the plot of a video that was watched by the user and a large amount of plots that is stored in a movie database. Since it is independent from the number of user ratings, it is able to propose famous and beloved movies as well as old or unheard movies/programs that are still strongly related to the content of the video the user has watched. We experimented different methodologies to compare natural language descriptions of movies (plots) and evaluated the Latent Semantic Analysis (LSA) to be the superior one in supporting the selection of similar plots. In order to increase the efficiency of LSA, different models have been experimented and in the end, a recommendation system that is able to compare about two hundred thousands movie plots in less than a minute has been developed
Transcriptional landscape of the human and fly genomes: Nonlinear and multifunctional modular model of transcriptomes
Regions of the genome not coding for proteins or not involved in cis-acting regulatory activities are frequently viewed as lacking in functional value. However, a number of recent large-scale studies have revealed significant regulated transcription of unannotated portions of a variety of plant and animal genomes, allowing a new appreciation of the widespread transcription of large portions of the genome. High-resolution mapping of the sites of transcription of the human and fly genomes has provided an alternative picture of the extent and organization of transcription and has offered insights for biological functions of some of the newly identified unannotated transcripts. Considerable portions of the unannotated transcription observed are developmental or cell-type-specific parts of protein-coding transcripts, often serving as novel, alternative 5′ transcriptional start sites. These distal 5′ portions are often situated at significant distances from the annotated gene and alternatively join with or ignore portions of other intervening genes to comprise novel unannotated protein-coding transcripts. These data support an interlaced model of the genome in which many regions serve multifunctional purposes and are highly modular in their utilization. This model illustrates the underappreciated organizational complexity of the genome and one of the functional roles of transcription from unannotated portions of the genome. Copyright 2006, Cold Spring Harbor Laboratory Press © 2006 Cold Spring Harbor Laboratory Press
Gene Function Classification Using Bayesian Models with Hierarchy-Based Priors
We investigate the application of hierarchical classification schemes to the
annotation of gene function based on several characteristics of protein
sequences including phylogenic descriptors, sequence based attributes, and
predicted secondary structure. We discuss three Bayesian models and compare
their performance in terms of predictive accuracy. These models are the
ordinary multinomial logit (MNL) model, a hierarchical model based on a set of
nested MNL models, and a MNL model with a prior that introduces correlations
between the parameters for classes that are nearby in the hierarchy. We also
provide a new scheme for combining different sources of information. We use
these models to predict the functional class of Open Reading Frames (ORFs) from
the E. coli genome. The results from all three models show substantial
improvement over previous methods, which were based on the C5 algorithm. The
MNL model using a prior based on the hierarchy outperforms both the
non-hierarchical MNL model and the nested MNL model. In contrast to previous
attempts at combining these sources of information, our approach results in a
higher accuracy rate when compared to models that use each data source alone.
Together, these results show that gene function can be predicted with higher
accuracy than previously achieved, using Bayesian models that incorporate
suitable prior information
Computational Indistinguishability between Quantum States and Its Cryptographic Application
We introduce a computational problem of distinguishing between two specific
quantum states as a new cryptographic problem to design a quantum cryptographic
scheme that is "secure" against any polynomial-time quantum adversary. Our
problem, QSCDff, is to distinguish between two types of random coset states
with a hidden permutation over the symmetric group of finite degree. This
naturally generalizes the commonly-used distinction problem between two
probability distributions in computational cryptography. As our major
contribution, we show that QSCDff has three properties of cryptographic
interest: (i) QSCDff has a trapdoor; (ii) the average-case hardness of QSCDff
coincides with its worst-case hardness; and (iii) QSCDff is computationally at
least as hard as the graph automorphism problem in the worst case. These
cryptographic properties enable us to construct a quantum public-key
cryptosystem, which is likely to withstand any chosen plaintext attack of a
polynomial-time quantum adversary. We further discuss a generalization of
QSCDff, called QSCDcyc, and introduce a multi-bit encryption scheme that relies
on similar cryptographic properties of QSCDcyc.Comment: 24 pages, 2 figures. We improved presentation, and added more detail
proofs and follow-up of recent wor
Machine Learning in Automated Text Categorization
The automated categorization (or classification) of texts into predefined
categories has witnessed a booming interest in the last ten years, due to the
increased availability of documents in digital form and the ensuing need to
organize them. In the research community the dominant approach to this problem
is based on machine learning techniques: a general inductive process
automatically builds a classifier by learning, from a set of preclassified
documents, the characteristics of the categories. The advantages of this
approach over the knowledge engineering approach (consisting in the manual
definition of a classifier by domain experts) are a very good effectiveness,
considerable savings in terms of expert manpower, and straightforward
portability to different domains. This survey discusses the main approaches to
text categorization that fall within the machine learning paradigm. We will
discuss in detail issues pertaining to three different problems, namely
document representation, classifier construction, and classifier evaluation.Comment: Accepted for publication on ACM Computing Survey
Time-Sensitive User Profile for Optimizing Search Personlization
International audienceThanks to social Web services, Web search engines have the opportunity to afford personalized search results that better fit the user’s information needs and interests. To achieve this goal, many personalized search approaches explore user’s social Web interactions to extract his preferences and interests, and use them to model his profile. In our approach, the user profile is implicitly represented as a vector of weighted terms which correspond to the user’s interests extracted from his online social activities. As the user interests may change over time, we propose to weight profiles terms not only according to the content of these activities but also by considering the freshness. More precisely, the weights are adjusted with a temporal feature. In order to evaluate our approach, we model the user profile according to data collected from Twitter. Then, we rerank initial search results accurately to the user profile. Moreover, we proved the significance of adding a temporal feature by comparing our method with baselines models that does not consider the user profile dynamics
- …