565,732 research outputs found
Mutual learning in a tree parity machine and its application to cryptography
Mutual learning of a pair of tree parity machines with continuous and
discrete weight vectors is studied analytically. The analysis is based on a
mapping procedure that maps the mutual learning in tree parity machines onto
mutual learning in noisy perceptrons. The stationary solution of the mutual
learning in the case of continuous tree parity machines depends on the learning
rate where a phase transition from partial to full synchronization is observed.
In the discrete case the learning process is based on a finite increment and a
full synchronized state is achieved in a finite number of steps. The
synchronization of discrete parity machines is introduced in order to construct
an ephemeral key-exchange protocol. The dynamic learning of a third tree parity
machine (an attacker) that tries to imitate one of the two machines while the
two still update their weight vectors is also analyzed. In particular, the
synchronization times of the naive attacker and the flipping attacker recently
introduced in [1] are analyzed. All analytical results are found to be in good
agreement with simulation results
Adaptation of the Electric Machines Learning Process to the European Higher, Education Area
In this paper the basic lines of a complete teaching methodology that has been developed to adaptthe electric machines learning process to the European Higher Education Area (EHEA) arepresented. New teaching materials that are specific to Electric Machines have been created(textbooks, self-learning e-books, guidelines for achieving teamwork research, etc.). Working ingroups has been promoted, as well as problem solving and self-learning exercises, all of which areevaluated in a way that encourages students' participation. Finally, the students' learning process inthe lab has been improved by the development both of a new methodology to follow in the lab andnew workbenches with industrial machines that are easier to use and also enable the labexperiments to be automated. Finally, the first results obtained as a result of applying the proposedmethodology are presented
Stochastic Synapses Enable Efficient Brain-Inspired Learning Machines
Recent studies have shown that synaptic unreliability is a robust and
sufficient mechanism for inducing the stochasticity observed in cortex. Here,
we introduce Synaptic Sampling Machines, a class of neural network models that
uses synaptic stochasticity as a means to Monte Carlo sampling and unsupervised
learning. Similar to the original formulation of Boltzmann machines, these
models can be viewed as a stochastic counterpart of Hopfield networks, but
where stochasticity is induced by a random mask over the connections. Synaptic
stochasticity plays the dual role of an efficient mechanism for sampling, and a
regularizer during learning akin to DropConnect. A local synaptic plasticity
rule implementing an event-driven form of contrastive divergence enables the
learning of generative models in an on-line fashion. Synaptic sampling machines
perform equally well using discrete-timed artificial units (as in Hopfield
networks) or continuous-timed leaky integrate & fire neurons. The learned
representations are remarkably sparse and robust to reductions in bit precision
and synapse pruning: removal of more than 75% of the weakest connections
followed by cursory re-learning causes a negligible performance loss on
benchmark classification tasks. The spiking neuron-based synaptic sampling
machines outperform existing spike-based unsupervised learners, while
potentially offering substantial advantages in terms of power and complexity,
and are thus promising models for on-line learning in brain-inspired hardware
- …