104,353 research outputs found

    A Comprehensive Analysis of Time Series Segmentation on the Japanese Stock Prices

    Full text link
    This study conducts a comprehensive analysis of time series segmentation on the Japanese stock prices listed on the first section of the Tokyo Stock Exchange during the period from 4 January 2000 to 30 January 2012. A recursive segmentation procedure is used under the assumption of a Gaussian mixture. The daily number of each quintile of volatilities for all the segments is investigated empirically. It is found that from June 2004 to June 2007, a large majority of stocks are stable and that from 2008 several stocks showed instability. On March 2011, the daily number of instable securities steeply increased due to societal turmoil influenced by the East Japan Great Earthquake. It is concluded that the number of stocks included in each quintile of volatilities provides useful information on macroeconomic situations.Comment: 10 pages, 5 figures, submitted to the 4th World Congress on Social Simulation (WCSS2012

    Why Model?

    Get PDF
    This address treats some enduring misconceptions about modeling. One of these is that the goal is always prediction. The lecture distinguishes between explanation and prediction as modeling goals, and offers sixteen reasons other than prediction to build a model. It also challenges the common assumption that scientific theories arise from and 'summarize' data, when often, theories precede and guide data collection; without theory, in other words, it is not clear what data to collect. Among other things, it also argues that the modeling enterprise enforces habits of mind essential to freedom. It is based on the author's 2008 Bastille Day keynote address to the Second World Congress on Social Simulation, George Mason University, and earlier addresses at the Institute of Medicine, the University of Michigan, and the Santa Fe Institute.[No keywords]

    Parameter Sensitivity Analysis of Social Spider Algorithm

    Full text link
    Social Spider Algorithm (SSA) is a recently proposed general-purpose real-parameter metaheuristic designed to solve global numerical optimization problems. This work systematically benchmarks SSA on a suite of 11 functions with different control parameters. We conduct parameter sensitivity analysis of SSA using advanced non-parametric statistical tests to generate statistically significant conclusion on the best performing parameter settings. The conclusion can be adopted in future work to reduce the effort in parameter tuning. In addition, we perform a success rate test to reveal the impact of the control parameters on the convergence speed of the algorithm

    Water and energy-based optimisation of a “MiniCity”: A system dynamics approach

    Get PDF

    The nature of evidence: how well do 'facts' travel? Annual report 2005-2006

    Get PDF
    corecore