1,339 research outputs found
Confinement induced instability of thin elastic film
A confined incompressible elastic film does not deform uniformly when
subjected to adhesive interfacial stresses but with undulations which have a
characteristic wavelength scaling linearly with the thickness of the film. In
the classical peel geometry, undulations appear along the contact line below a
critical film thickness or below a critical curvature of the plate.
Perturbation analysis of the stress equilibrium equations shows that for a
critically confined film the total excess energy indeed attains a minima for a
finite amplitude of the perturbations which grow with further increase in the
confinement.Comment: 11 pages, 6 figure
Friction and abrasion of elastomeric materials
An abrasion apparatus is described. Experimental measurements are reported for four representative elastomeric materials, including a typical high-quality tire tread material and a possible replacement material for aircraft tire treads based on transpolypentenamer (TPPR). Measurements are carried out at different levels of frictional work input, corresponding to different severities of wear, and at both ambient temperature and at 100 C. Results indicate the marked superiority in abrasion resistance of the material based on TPPR, especially at 100 C, in comparison with the other materials examined
Magic angles and cross-hatching instability in hydrogel fracture
The full 2D analysis of roughness profiles of fracture surfaces resulting
from quasi-static crack propagation in gelatin gels reveals an original
behavior characterized by (i) strong anisotropy with maximum roughness at
-independent symmetry-preserving angles, (ii) a sub-critical instability
leading, below a critical velocity, to a cross-hatched regime due to straight
macrosteps drifting at the same magic angles and nucleated on crack-pinning
network inhomogeneities. Step height values are determined by the width of the
strain-hardened zone, governed by the elastic crack blunting characteristic of
soft solids with breaking stresses much larger that low strain moduli
A pearl on SAT solving in Prolog
A succinct SAT solver is presented that exploits the control provided by delay declarations to implement watched literals and unit propagation. Despite its brevity the solver is surprisingly powerful and its elegant use of Prolog constructs is presented as a programming pearl
Bone Proteomics Method Optimization for Forensic Investigations
\ua9 2024 The Authors. Published by American Chemical Society.The application of proteomic analysis to forensic skeletal remains has gained significant interest in improving biological and chronological estimations in medico-legal investigations. To enhance the applicability of these analyses to forensic casework, it is crucial to maximize throughput and proteome recovery while minimizing interoperator variability and laboratory-induced post-translational protein modifications (PTMs). This work compared different workflows for extracting, purifying, and analyzing bone proteins using liquid chromatography with tandem mass spectrometry (LC-MS)/MS including an in-StageTip protocol previously optimized for forensic applications and two protocols using novel suspension-trap technology (S-Trap) and different lysis solutions. This study also compared data-dependent acquisition (DDA) with data-independent acquisition (DIA). By testing all of the workflows on 30 human cortical tibiae samples, S-Trap workflows resulted in increased proteome recovery with both lysis solutions tested and in decreased levels of induced deamidations, and the DIA mode resulted in greater sensitivity and window of identification for the identification of lower-abundance proteins, especially when open-source software was utilized for data processing in both modes. The newly developed S-Trap protocol is, therefore, suitable for forensic bone proteomic workflows and, particularly when paired with DIA mode, can offer improved proteomic outcomes and increased reproducibility, showcasing its potential in forensic proteomics and contributing to achieving standardization in bone proteomic analyses for forensic applications
Certainty Closure: Reliable Constraint Reasoning with Incomplete or Erroneous Data
Constraint Programming (CP) has proved an effective paradigm to model and
solve difficult combinatorial satisfaction and optimisation problems from
disparate domains. Many such problems arising from the commercial world are
permeated by data uncertainty. Existing CP approaches that accommodate
uncertainty are less suited to uncertainty arising due to incomplete and
erroneous data, because they do not build reliable models and solutions
guaranteed to address the user's genuine problem as she perceives it. Other
fields such as reliable computation offer combinations of models and associated
methods to handle these types of uncertain data, but lack an expressive
framework characterising the resolution methodology independently of the model.
We present a unifying framework that extends the CP formalism in both model
and solutions, to tackle ill-defined combinatorial problems with incomplete or
erroneous data. The certainty closure framework brings together modelling and
solving methodologies from different fields into the CP paradigm to provide
reliable and efficient approches for uncertain constraint problems. We
demonstrate the applicability of the framework on a case study in network
diagnosis. We define resolution forms that give generic templates, and their
associated operational semantics, to derive practical solution methods for
reliable solutions.Comment: Revised versio
New scaling for the alpha effect in slowly rotating turbulence
Using simulations of slowly rotating stratified turbulence, we show that the
alpha effect responsible for the generation of astrophysical magnetic fields is
proportional to the logarithmic gradient of kinetic energy density rather than
that of momentum, as was previously thought. This result is in agreement with a
new analytic theory developed in this paper for large Reynolds numbers. Thus,
the contribution of density stratification is less important than that of
turbulent velocity. The alpha effect and other turbulent transport coefficients
are determined by means of the test-field method. In addition to forced
turbulence, we also investigate supernova-driven turbulence and stellar
convection. In some cases (intermediate rotation rate for forced turbulence,
convection with intermediate temperature stratification, and supernova-driven
turbulence) we find that the contribution of density stratification might be
even less important than suggested by the analytic theory.Comment: 10 pages, 9 figures, revised version, Astrophys. J., in pres
Phase transition for cutting-plane approach to vertex-cover problem
We study the vertex-cover problem which is an NP-hard optimization problem
and a prototypical model exhibiting phase transitions on random graphs, e.g.,
Erdoes-Renyi (ER) random graphs. These phase transitions coincide with changes
of the solution space structure, e.g, for the ER ensemble at connectivity
c=e=2.7183 from replica symmetric to replica-symmetry broken. For the
vertex-cover problem, also the typical complexity of exact branch-and-bound
algorithms, which proceed by exploring the landscape of feasible
configurations, change close to this phase transition from "easy" to "hard". In
this work, we consider an algorithm which has a completely different strategy:
The problem is mapped onto a linear programming problem augmented by a
cutting-plane approach, hence the algorithm operates in a space OUTSIDE the
space of feasible configurations until the final step, where a solution is
found. Here we show that this type of algorithm also exhibits an "easy-hard"
transition around c=e, which strongly indicates that the typical hardness of a
problem is fundamental to the problem and not due to a specific representation
of the problem.Comment: 4 pages, 3 figure
Optimization by Quantum Annealing: Lessons from hard 3-SAT cases
The Path Integral Monte Carlo simulated Quantum Annealing algorithm is
applied to the optimization of a large hard instance of the Random 3-SAT
Problem (N=10000). The dynamical behavior of the quantum and the classical
annealing are compared, showing important qualitative differences in the way of
exploring the complex energy landscape of the combinatorial optimization
problem. At variance with the results obtained for the Ising spin glass and for
the Traveling Salesman Problem, in the present case the linear-schedule Quantum
Annealing performance is definitely worse than Classical Annealing.
Nevertheless, a quantum cooling protocol based on field-cycling and able to
outperform standard classical simulated annealing over short time scales is
introduced.Comment: 10 pages, 6 figures, submitted to PR
Highlights of the Zeno Results from the USMP-2 Mission
The Zeno instrument, a High-precision, light-scattering spectrometer, was built to measure the decay rates of density fluctuations in xenon near its liquid-vapor critical point in the low-gravity environment of the U.S. Space Shuttle. Eliminating the severe density gradients created in a critical fluid by Earth's gravity, we were able to make measurements to within 100 microKelvin of the critical point. The instrument flew for fourteen days in March, 1994 on the Space Shuttle Columbia, STS-62 flight, as part of the very successful USMP-2 payload. We describe the instrument and document its performance on orbit, showing that it comfortably reached the desired 3 microKelvin temperature control of the sample. Locating the critical temperature of the sample on orbit was a scientific challenge; we discuss the advantages and short-comings of the two techniques we used. Finally we discuss problems encountered with making measurements of the turbidity of the sample, and close with the results of the measurement of the decay rates of the critical-point fluctuations
- …