2,187 research outputs found

    A gradient system with a wiggly energy and relaxed EDP-convergence

    Get PDF
    If gradient systems depend on a microstructure, we want to derive a macroscopic gradient structure describing the effective behavior of the microscopic effects. We introduce a notion of evolutionary Gamma-convergence that relates the microscopic energy and the microscopic dissipation potential with their macroscopic limits via Gamma-convergence. This new notion generalizes the concept of EDP-convergence, which was introduced in arXiv:1507.06322, and is called "relaxed EDP-convergence". Both notions are based on De Giorgi's energy-dissipation principle, however the special structure of the dissipation functional in terms of the primal and dual dissipation potential is, in general, not preserved under Gamma-convergence. By investigating the kinetic relation directly and using general forcings we still derive a unique macroscopic dissipation potential. The wiggly-energy model of James et al serves as a prototypical example where this nontrivial limit passage can be fully analyzed.Comment: 43 pages, 8 figure

    Shape optimization problems on metric measure spaces

    Get PDF
    We consider shape optimization problems of the form min{J(Ω) : ΩX, m(Ω)c},\min\big\{J(\Omega)\ :\ \Omega\subset X,\ m(\Omega)\le c\big\}, where XX is a metric measure space and JJ is a suitable shape functional. We adapt the notions of γ\gamma-convergence and weak γ\gamma-convergence to this new general abstract setting to prove the existence of an optimal domain. Several examples are pointed out and discussed.Comment: 27 pages, the final publication is available at http://www.journals.elsevier.com/journal-of-functional-analysis

    Pointwise Convergence in Probability of General Smoothing Splines

    Get PDF
    Establishing the convergence of splines can be cast as a variational problem which is amenable to a Γ\Gamma-convergence approach. We consider the case in which the regularization coefficient scales with the number of observations, nn, as λn=np\lambda_n=n^{-p}. Using standard theorems from the Γ\Gamma-convergence literature, we prove that the general spline model is consistent in that estimators converge in a sense slightly weaker than weak convergence in probability for p12p\leq \frac{1}{2}. Without further assumptions we show this rate is sharp. This differs from rates for strong convergence using Hilbert scales where one can often choose p>12p>\frac{1}{2}
    corecore