25,321 research outputs found
Memristor models for machine learning
In the quest for alternatives to traditional CMOS, it is being suggested that
digital computing efficiency and power can be improved by matching the
precision to the application. Many applications do not need the high precision
that is being used today. In particular, large gains in area- and power
efficiency could be achieved by dedicated analog realizations of approximate
computing engines. In this work, we explore the use of memristor networks for
analog approximate computation, based on a machine learning framework called
reservoir computing. Most experimental investigations on the dynamics of
memristors focus on their nonvolatile behavior. Hence, the volatility that is
present in the developed technologies is usually unwanted and it is not
included in simulation models. In contrast, in reservoir computing, volatility
is not only desirable but necessary. Therefore, in this work, we propose two
different ways to incorporate it into memristor simulation models. The first is
an extension of Strukov's model and the second is an equivalent Wiener model
approximation. We analyze and compare the dynamical properties of these models
and discuss their implications for the memory and the nonlinear processing
capacity of memristor networks. Our results indicate that device variability,
increasingly causing problems in traditional computer design, is an asset in
the context of reservoir computing. We conclude that, although both models
could lead to useful memristor based reservoir computing systems, their
computational performance will differ. Therefore, experimental modeling
research is required for the development of accurate volatile memristor models.Comment: 4 figures, no tables. Submitted to neural computatio
On the Solution of Linear Programming Problems in the Age of Big Data
The Big Data phenomenon has spawned large-scale linear programming problems.
In many cases, these problems are non-stationary. In this paper, we describe a
new scalable algorithm called NSLP for solving high-dimensional, non-stationary
linear programming problems on modern cluster computing systems. The algorithm
consists of two phases: Quest and Targeting. The Quest phase calculates a
solution of the system of inequalities defining the constraint system of the
linear programming problem under the condition of dynamic changes in input
data. To this end, the apparatus of Fejer mappings is used. The Targeting phase
forms a special system of points having the shape of an n-dimensional
axisymmetric cross. The cross moves in the n-dimensional space in such a way
that the solution of the linear programming problem is located all the time in
an "-vicinity of the central point of the cross.Comment: Parallel Computational Technologies - 11th International Conference,
PCT 2017, Kazan, Russia, April 3-7, 2017, Proceedings (to be published in
Communications in Computer and Information Science, vol. 753
XBioSiP: A Methodology for Approximate Bio-Signal Processing at the Edge
Bio-signals exhibit high redundancy, and the algorithms for their processing
are inherently error resilient. This property can be leveraged to improve the
energy-efficiency of IoT-Edge (wearables) through the emerging trend of
approximate computing. This paper presents XBioSiP, a novel methodology for
approximate bio-signal processing that employs two quality evaluation stages,
during the pre-processing and bio-signal processing stages, to determine the
approximation parameters. It thereby achieves high energy savings while
satisfying the user-determined quality constraint. Our methodology achieves, up
to 19x and 22x reduction in the energy consumption of a QRS peak detection
algorithm for 0% and <1% loss in peak detection accuracy, respectively.Comment: Accepted for publication at the Design Automation Conference 2019
(DAC'19), Las Vegas, Nevada, US
Quantum machine learning: a classical perspective
Recently, increased computational power and data availability, as well as
algorithmic advances, have led machine learning techniques to impressive
results in regression, classification, data-generation and reinforcement
learning tasks. Despite these successes, the proximity to the physical limits
of chip fabrication alongside the increasing size of datasets are motivating a
growing number of researchers to explore the possibility of harnessing the
power of quantum computation to speed-up classical machine learning algorithms.
Here we review the literature in quantum machine learning and discuss
perspectives for a mixed readership of classical machine learning and quantum
computation experts. Particular emphasis will be placed on clarifying the
limitations of quantum algorithms, how they compare with their best classical
counterparts and why quantum resources are expected to provide advantages for
learning problems. Learning in the presence of noise and certain
computationally hard problems in machine learning are identified as promising
directions for the field. Practical questions, like how to upload classical
data into quantum form, will also be addressed.Comment: v3 33 pages; typos corrected and references adde
- …