5,573 research outputs found

    Comment on "Why quantum mechanics cannot be formulated as a Markov process"

    Get PDF
    In the paper with the above title, D. T. Gillespie [Phys. Rev. A 49, 1607, (1994)] claims that the theory of Markov stochastic processes cannot provide an adequate mathematical framework for quantum mechanics. In conjunction with the specific quantum dynamics considered there, we give a general analysis of the associated dichotomic jump processes. If we assume that Gillespie's "measurement probabilities" \it are \rm the transition probabilities of a stochastic process, then the process must have an invariant (time independent) probability measure. Alternatively, if we demand the probability measure of the process to follow the quantally implemented (via the Born statistical postulate) evolution, then we arrive at the jump process which \it can \rm be interpreted as a Markov process if restricted to a suitable duration time. However, there is no corresponding Markov process consistent with the Z2Z_2 event space assumption, if we require its existence for all times t∈R+t\in R_+.Comment: Latex file, resubm. to Phys. Rev.

    Shell Model of Two-dimensional Turbulence in Polymer Solutions

    Full text link
    We address the effect of polymer additives on two dimensional turbulence, an issue that was studied recently in experiments and direct numerical simulations. We show that the same simple shell model that reproduced drag reduction in three-dimensional turbulence reproduces all the reported effects in the two-dimensional case. The simplicity of the model offers a straightforward understanding of the all the major effects under consideration

    Large-scale albuminuria screen for nephropathy models in chemically induced mouse mutants

    Get PDF
    Background/Aim: Phenotype-driven screening of a great pool of randomly mutant mice and subsequent selection of animals showing symptoms equivalent to human kidney diseases may result in the generation of novel suitable models for the study of the pathomechanisms and the identification of genes involved in kidney dysfunction. Methods: We carried out a large-scale analysis of ethylnitrosourea (ENU)-induced mouse mutants for albuminuria by using qualitative SDS-polyacrylamide gel electrophoresis. Results: The primary albuminuria screen preceded the comprehensive phenotypic mutation analysis in a part of the mice of the Munich ENU project to avoid loss of mutant animals as a consequence of prolonged suffering from severe nephropathy. The primary screen detected six confirmed phenotypic variants in 2,011 G1 animals screened for dominant mutations and no variant in 48 G3 pedigrees screened for recessive mutations. Further breeding experiments resulted in two lines showing a low phenotypic penetrance of albuminuria. The secondary albuminuria screen was carried out in mutant lines which were established in the Munich ENU project without preceding primary albuminuria analysis. Two lines showing increased plasma urea levels were chosen to clarify if severe kidney lesions are involved in the abnormal phenotype. This analysis revealed severe albuminuria in mice which are affected by a recessive mutation leading to increased plasma urea and cholesterol levels. Conclusion: Thus, the phenotypic selection of ENU-induced mutants according to the parameter proteinuria in principle demonstrates the feasibility to identify nephropathy phenotypes in ENU-mutagenized mice. Copyright (C) 2005 S. Karger AG, Basel

    Weak nuclear forces cause the strong nuclear force

    Get PDF
    We determine the strength of the weak nuclear force which holds the lattices of the elementary particles together. We also determine the strength of the strong nuclear force which emanates from the sides of the nuclear lattices. The strong force is the sum of the unsaturated weak forces at the surface of the nuclear lattices. The strong force is then about ten to the power of 6 times stronger than the weak force between two lattice points.Comment: 12 pages, 1 figur

    Exact Ground State and Finite Size Scaling in a Supersymmetric Lattice Model

    Full text link
    We study a model of strongly correlated fermions in one dimension with extended N=2 supersymmetry. The model is related to the spin S=1/2S=1/2 XXZ Heisenberg chain at anisotropy Δ=−1/2\Delta=-1/2 with a real magnetic field on the boundary. We exploit the combinatorial properties of the ground state to determine its exact wave function on finite lattices with up to 30 sites. We compute several correlation functions of the fermionic and spin fields. We discuss the continuum limit by constructing lattice observables with well defined finite size scaling behavior. For the fermionic model with periodic boundary conditions we give the emptiness formation probability in closed form.Comment: 4 pages, 4 eps figure

    The Speed of Light and the Hubble Parameter: The Mass-Boom Effect

    Get PDF
    We prove here that Newtons universal gravitation and momentum conservation laws together reproduce Weinbergs relation. It is shown that the Hubble parameter H must be built in this relation, or equivalently the age of the Universe t. Using a wave-to-particle interaction technique we then prove that the speed of light c decreases with cosmological time, and that c is proportional to the Hubble parameter H. We see the expansion of the Universe as a local effect due to the LAB value of the speed of light co taken as constant. We present a generalized red shift law and find a predicted acceleration for photons that agrees well with the result from Pioneer 10/11 anomalous acceleration. We finally present a cosmological model coherent with the above results that we call the Mass-Boom. It has a linear increase of mass m with time as a result of the speed of light c linear decrease with time, and the conservation of momentum mc. We obtain the baryonic mass parameter equal to the curvature parameter, omega m = omega k, so that the model is of the type of the Einstein static, closed, finite, spherical, unlimited, with zero cosmological constant. This model is the cosmological view as seen by photons, neutrinos, tachyons etc. in contrast with the local view, the LAB reference. Neither dark matter nor dark energy is required by this model. With an initial constant speed of light during a short time we get inflation (an exponential expansion). This converts, during the inflation time, the Plancks fluctuation length of 10-33 cm to the present size of the Universe (about 1028 cm, constant from then on). Thereafter the Mass-Boom takes care to bring the initial values of the Universe (about 1015 gr) to the value at the present time of about 1055 gr.Comment: 15 pages, presented at the 9th Symposium on "Frontiers of Fundamental Physics", 7-9 Jan. 2008, University of Udine, Italy. Changed content

    Hidden-variable theory versus Copenhagen quantum mechanics

    Full text link
    The main assumptions the Copenhagen quantum mechanics has been based on will be summarized and the known (not yet decided) contradiction between Einstein and Bohr will be newly analyzed. The given assumptions have been represented basically by time-dependent Schroedinger equation, to which some further assumptions have been added. Some critical comments have been raised against the given mathematical model structure by Pauli (1933) and by Susskind and Glogover (1964). They may be removed if only the Schroedinger equation is conserved and the additional assumptions are abandoned, as shown recently. It seems to be in contradiction to the numerous declarations that the Copenhagen model has been approved by experimental results. However, in the most of these experiments only the agreement with the mere Schroedinger equation has been tested. All mentioned assumptions have been tested practically only in the EPR experiment (measurement of coincidence light transmission through two polarizers) proposed originally by Einstein (1935). Also these experimental results have been interpreted as supporting the Copenhagen alternative, which has not been, however, true. In fact the microscopic world may be described correspondingly only with the help of the hidden-variable theory that is represented by the Schroedinger equation without mentioned additional assumptions, which has the consequence that the earlier interpretation gap between microscopic and macroscopic worlds has been removed. The only difference concerns the existence of discrete states. The possibilities of the human reason of getting to know the nature will be also shortly discussed in the beginning of this contribution.Comment: 10 pages, 2 figures; v2: local refinements and improvements of the tex

    Biological Principles in Self-Organization of Young Brain - Viewed from Kohonen Model

    Get PDF
    Variants of the Kohonen model are proposed to study biological principles of self-organization in a model of young brain. We suggest a function to measure aquired knowledge and use it to auto-adapt the topology of neuronal connectivity, yielding substantial organizational improvement relative to the standard model. In the early phase of organization with most intense learning, we observe that neural connectivity is of Small World type, which is very efficient to organize neurons in response to stimuli. In analogy to human brain where pruning of neural connectivity (and neuron cell death) occurs in early life, this feature is present also in our model, which is found to stabilize neuronal response to stimuli
    • 

    corecore