30,399 research outputs found
ShenZhen transportation system (SZTS): a novel big data benchmark suite
Data analytics is at the core of the supply chain for both products and services in modern economies and societies. Big data workloads, however, are placing unprecedented demands on computing technologies, calling for a deep understanding and characterization of these emerging workloads. In this paper, we propose ShenZhen Transportation System (SZTS), a novel big data Hadoop benchmark suite comprised of real-life transportation analysis applications with real-life input data sets from Shenzhen in China. SZTS uniquely focuses on a specific and real-life application domain whereas other existing Hadoop benchmark suites, such as HiBench and CloudRank-D, consist of generic algorithms with synthetic inputs. We perform a cross-layer workload characterization at the microarchitecture level, the operating system (OS) level, and the job level, revealing unique characteristics of SZTS compared to existing Hadoop benchmarks as well as general-purpose multi-core PARSEC benchmarks. We also study the sensitivity of workload behavior with respect to input data size, and we propose a methodology for identifying representative input data sets
Chaos in computer performance
Modern computer microprocessors are composed of hundreds of millions of
transistors that interact through intricate protocols. Their performance during
program execution may be highly variable and present aperiodic oscillations. In
this paper, we apply current nonlinear time series analysis techniques to the
performances of modern microprocessors during the execution of prototypical
programs. Our results present pieces of evidence strongly supporting that the
high variability of the performance dynamics during the execution of several
programs display low-dimensional deterministic chaos, with sensitivity to
initial conditions comparable to textbook models. Taken together, these results
show that the instantaneous performances of modern microprocessors constitute a
complex (or at least complicated) system and would benefit from analysis with
modern tools of nonlinear and complexity science
Information geometric methods for complexity
Research on the use of information geometry (IG) in modern physics has
witnessed significant advances recently. In this review article, we report on
the utilization of IG methods to define measures of complexity in both
classical and, whenever available, quantum physical settings. A paradigmatic
example of a dramatic change in complexity is given by phase transitions (PTs).
Hence we review both global and local aspects of PTs described in terms of the
scalar curvature of the parameter manifold and the components of the metric
tensor, respectively. We also report on the behavior of geodesic paths on the
parameter manifold used to gain insight into the dynamics of PTs. Going
further, we survey measures of complexity arising in the geometric framework.
In particular, we quantify complexity of networks in terms of the Riemannian
volume of the parameter space of a statistical manifold associated with a given
network. We are also concerned with complexity measures that account for the
interactions of a given number of parts of a system that cannot be described in
terms of a smaller number of parts of the system. Finally, we investigate
complexity measures of entropic motion on curved statistical manifolds that
arise from a probabilistic description of physical systems in the presence of
limited information. The Kullback-Leibler divergence, the distance to an
exponential family and volumes of curved parameter manifolds, are examples of
essential IG notions exploited in our discussion of complexity. We conclude by
discussing strengths, limits, and possible future applications of IG methods to
the physics of complexity.Comment: review article, 60 pages, no figure
- …