12,818 research outputs found

    Status and Future Perspectives for Lattice Gauge Theory Calculations to the Exascale and Beyond

    Full text link
    In this and a set of companion whitepapers, the USQCD Collaboration lays out a program of science and computing for lattice gauge theory. These whitepapers describe how calculation using lattice QCD (and other gauge theories) can aid the interpretation of ongoing and upcoming experiments in particle and nuclear physics, as well as inspire new ones.Comment: 44 pages. 1 of USQCD whitepapers

    Bridging Proper Orthogonal Decomposition methods and augmented Newton-Krylov algorithms: an adaptive model order reduction for highly nonlinear mechanical problems

    Get PDF
    This article describes a bridge between POD-based model order reduction techniques and the classical Newton/Krylov solvers. This bridge is used to derive an efficient algorithm to correct, "on-the-fly", the reduced order modelling of highly nonlinear problems undergoing strong topological changes. Damage initiation problems are addressed and tackle via a corrected hyperreduction method. It is shown that the relevancy of reduced order model can be significantly improved with reasonable additional costs when using this algorithm, even when strong topological changes are involved

    Understanding the Role of Adaptivity in Machine Teaching: The Case of Version Space Learners

    Get PDF
    In real-world applications of education, an effective teacher adaptively chooses the next example to teach based on the learner's current state. However, most existing work in algorithmic machine teaching focuses on the batch setting, where adaptivity plays no role. In this paper, we study the case of teaching consistent, version space learners in an interactive setting. At any time step, the teacher provides an example, the learner performs an update, and the teacher observes the learner's new state. We highlight that adaptivity does not speed up the teaching process when considering existing models of version space learners, such as "worst-case" (the learner picks the next hypothesis randomly from the version space) and "preference-based" (the learner picks hypothesis according to some global preference). Inspired by human teaching, we propose a new model where the learner picks hypotheses according to some local preference defined by the current hypothesis. We show that our model exhibits several desirable properties, e.g., adaptivity plays a key role, and the learner's transitions over hypotheses are smooth/interpretable. We develop efficient teaching algorithms and demonstrate our results via simulation and user studies.Comment: NeurIPS 2018 (extended version

    Status and challenges of simulations with dynamical fermions

    Full text link
    An overview over the current state of algorithms for dynamical fermion simulations is given. In particular some insight into the functioning of the determinant spitting techniques is discussed. The critical slowing down of the simulations towards the continuum limit and the role of the boundary conditions is also reviewed.Comment: 20 pages, 9 figures, plenary talk presented at the 30th International Symposium on Lattice Field Theory - Lattice 2012, June 24-29, 2012 Cairns, Australi

    Quantum annealing for systems of polynomial equations

    Full text link
    Numerous scientific and engineering applications require numerically solving systems of equations. Classically solving a general set of polynomial equations requires iterative solvers, while linear equations may be solved either by direct matrix inversion or iteratively with judicious preconditioning. However, the convergence of iterative algorithms is highly variable and depends, in part, on the condition number. We present a direct method for solving general systems of polynomial equations based on quantum annealing, and we validate this method using a system of second-order polynomial equations solved on a commercially available quantum annealer. We then demonstrate applications for linear regression, and discuss in more detail the scaling behavior for general systems of linear equations with respect to problem size, condition number, and search precision. Finally, we define an iterative annealing process and demonstrate its efficacy in solving a linear system to a tolerance of 10810^{-8}.Comment: 11 pages, 4 figures. Added example for a system of quadratic equations. Supporting code is available at https://github.com/cchang5/quantum_poly_solver . This is a post-peer-review, pre-copyedit version of an article published in Scientific Reports. The final authenticated version is available online at: https://www.nature.com/articles/s41598-019-46729-

    Fast and Accurate Coarsening Simulation with an Unconditionally Stable Time Step

    Full text link
    We present Cahn-Hilliard and Allen-Cahn numerical integration algorithms that are unconditionally stable and so provide significantly faster accuracy-controlled simulation. Our stability analysis is based on Eyre's theorem and unconditional von Neumann stability analysis, both of which we present. Numerical tests confirm the accuracy of the von Neumann approach, which is straightforward and should be widely applicable in phase-field modeling. We show that accuracy can be controlled with an unbounded time step Delta-t that grows with time t as Delta-t ~ t^alpha. We develop a classification scheme for the step exponent alpha and demonstrate that a class of simple linear algorithms gives alpha=1/3. For this class the speed up relative to a fixed time step grows with the linear size of the system as N/log N, and we estimate conservatively that an 8192^2 lattice can be integrated 300 times faster than with the Euler method.Comment: 14 pages, 6 figure
    corecore