6,228 research outputs found
Addendum to "Nonlinear quantum evolution with maximal entropy production"
The author calls attention to previous work with related results, which has
escaped scrutiny before the publication of the article "Nonlinear quantum
evolution with maximal entropy production", Phys.Rev.A63, 022105 (2001).Comment: RevTex-latex2e, 2pgs., no figs.; brief report to appear in the May
2001 issue of Phys.Rev.
A variational algorithm for the detection of line segments
In this paper we propose an algorithm for the detection of edges in images
that is based on topological asymptotic analysis. Motivated from the
Mumford--Shah functional, we consider a variational functional that penalizes
oscillations outside some approximate edge set, which we represent as the union
of a finite number of thin strips, the width of which is an order of magnitude
smaller than their length. In order to find a near optimal placement of these
strips, we compute an asymptotic expansion of the functional with respect to
the strip size. This expansion is then employed for defining a (topological)
gradient descent like minimization method. As opposed to a recently proposed
method by some of the authors, which uses coverings with balls, the usage of
strips includes some directional information into the method, which can be used
for obtaining finer edges and can also result in a reduction of computation
times
A Fast General-Purpose Clustering Algorithm Based on FPGAs for High-Throughput Data Processing
We present a fast general-purpose algorithm for high-throughput clustering of
data "with a two dimensional organization". The algorithm is designed to be
implemented with FPGAs or custom electronics. The key feature is a processing
time that scales linearly with the amount of data to be processed. This means
that clustering can be performed in pipeline with the readout, without
suffering from combinatorial delays due to looping multiple times through all
the data. This feature makes this algorithm especially well suited for problems
where the data has high density, e.g. in the case of tracking devices working
under high-luminosity condition such as those of LHC or Super-LHC. The
algorithm is organized in two steps: the first step (core) clusters the data;
the second step analyzes each cluster of data to extract the desired
information. The current algorithm is developed as a clustering device for
modern high-energy physics pixel detectors. However, the algorithm has much
broader field of applications. In fact, its core does not specifically rely on
the kind of data or detector it is working for, while the second step can and
should be tailored for a given application. Applications can thus be foreseen
to other detectors and other scientific fields ranging from HEP calorimeters to
medical imaging. An additional advantage of this two steps approach is that the
typical clustering related calculations (second step) are separated from the
combinatorial complications of clustering. This separation simplifies the
design of the second step and it enables it to perform sophisticated
calculations achieving online-quality in online applications. The algorithm is
general purpose in the sense that only minimal assumptions on the kind of
clustering to be performed are made.Comment: 11th Frontier Detectors For Frontier Physics conference (2009
Quantum thermodynamic Carnot and Otto-like cycles for a two-level system
From the thermodynamic equilibrium properties of a two-level system with
variable energy-level gap , and a careful distinction between the Gibbs
relation and the energy balance equation , we infer some important aspects of the
second law of thermodynamics and, contrary to a recent suggestion based on the
analysis of an Otto-like thermodynamic cycle between two values of of
a spin-1/2 system, we show that a quantum thermodynamic Carnot cycle, with the
celebrated optimal efficiency , is possible in
principle with no need of an infinite number of infinitesimal processes,
provided we cycle smoothly over at least three (in general four) values of
, and we change not only along the isoentropics, but also
along the isotherms, e.g., by use of the recently suggested maser-laser tandem
technique. We derive general bounds to the net-work to high-temperature-heat
ratio for a Carnot cycle and for the 'inscribed' Otto-like cycle, and represent
these cycles on useful thermodynamic diagrams.Comment: RevTex4, 4 pages, 1 figur
Thermodynamic analysis of turbulent combustion in a spark ignition engine. Experimental evidence
A method independent of physical modeling assumptions is presented to analyze high speed flame photography and cylinder pressure measurements from a transparent piston spark ignition research engine. The method involves defining characteristic quantities of the phenomena of flame propagation and combustion, and estimating their values from the experimental information. Using only the pressure information, the mass fraction curves are examined. An empirical burning law is presented which simulates such curves. Statistical data for the characteristics delay and burning angles which show that cycle to cycle fractional variations are of the same order of magnitude for both angles are discussed. The enflamed and burnt mass fractions are compared as are the rates of entrainment and burning
Electrons diffusion and signal noise contributions on electron clusters detection efficiency
The Cluster Counting (CC) technique, proposed for dE/dx measurements with the SuperB drift-chamber, could, significantly, improve particle identification by avoiding the fluctuations involved in charge measurements. As the technique is quite sensitive to the detector working conditions and to the front-end chain response, in this note we have investigated the effects on clusters detection efficiency of electron diffusion, preamplifier frequencyresponse and Signal-to-Noise Ratio (SNR) using different algorithms. The evaluation is based on Garfield datasets, generated for a single cell geometry, at different impact points for Ï /”/e particles with momenta 120, 140, 160, 180 and 210 MeV. The current waveforms generated by Garfield have been shaped according to the preamplifier response and different amounts of white gaussian noise has been added to the waveforms to simulate different SNRs. Finally an estimation of Ï /”/e separation is shown
Impianti di ionizzazione Cu-Ag per la di disinfezione delle acque e rischi di corrosione per spostamento
The copper and silver ionization system is one of the water sanitation treatments. Effective ionization occurs if the
content of copper ions in solution is 0.2-0.4 mg/L and that of silver ions is 0.02-0.04 mg/L. An excess of copper
and silver ions can react with other metal surfaces, triggering a deposition reaction, allowing the formation of
deposits of more noble metals, and then promoting a localized corrosion phenomenon due to galvanic coupling.
In the present paper, two case histories will be presented: one related to a legionella sanitizing plant system of a
hospital; the second related to a system of water purification of a vessel on a boat. In both cases, working
conditions are illustrated, corrosion morphology is described, focusing on the presence of copper and silver
deposits, and the cause of corrosion is presented, estimating a reliable corrosion rate
Use of degree of disequilibrium analysis to select kinetic constraints for the rate-controlled constrained-equilibrium (RCCE) method
The Rate-Controlled Constrained-Equilibrium (RCCE) method provides a general framework that enables, with the same ease, reduced order kinetic modelling at three different levels of approximation: shifting equilibrium, frozen equilibrium, as well as non-equilibrium chemical kinetics. The method in general requires a significantly smaller number of differential equations than the dimension of the underlying Detailed Kinetic Model (DKM) for acceptable accuracies. To provide accurate approximations, however, the method requires accurate identification of the bottleneck kinetic mechanisms responsible for slowing down the relaxation of the state of the system towards local chemical equilibrium. In other words, the method requires that such bottleneck mechanisms be characterized by means of a set of representative constraints. So far, a drawback of the RCCE method has been the absence of a systematic algorithm that would allow a fully automatable identification of the best constraints for a given range of thermodynamic conditions and a required level of approximation. In this paper, we provide the first of two steps towards such algorithm based on the analysis of the degrees of disequilibrium (DoD) of chemical reactions in the underlying DKM. In any given DKM the number of rate-limiting kinetic bottlenecks is generally much smaller than the number of species in the model. As a result, the DoDs of all the chemical reactions effectively assemble into a small number of groups that bear the information of the rate-controlling constraints. The DoDs of all reactions in each group exhibit almost identical behaviour (time evolution, spatial dependence). Upon identification of these groups, the proposed kernel analysis of N matrices that are obtained from the stoichiometric coefficients yields the N constraints that effectively control the dynamics of the system. The method is demonstrated within the framework of modeling the expansion of products of the oxy-combustion of hydrogen through a quasi one-dimensional supersonic nozzle. The analysis predicts and RCCE simulations confirm that, under the geometrical and boundary conditions considered, the underlying DKM is accurately represented by only two bottleneck kinetic mechanisms, instead of the three constraints identified for the same problem in a recently published work also based, in part, on DoD analysis
- âŠ