3,335,433 research outputs found

    Neutron Resonance Data Exclude Random Matrix Theory

    Full text link
    Almost since the time it was formulated, the overwhelming consensus has been that random matrix theory (RMT) is in excellent agreement with neutron resonance data. However, over the past few years, we have obtained new neutron-width data at Oak Ridge and Los Alamos National Laboratories that are in stark disagreement with this theory. We also have reanalyzed neutron widths in the most famous data set, the nuclear data ensemble (NDE), and found that it is seriously flawed, and, when analyzed carefully, excludes RMT with high confidence. More recently, we carefully examined energy spacings for these same resonances in the NDE using the Δ3\Delta_{3} statistic. We conclude that the data can be found to either confirm or refute the theory depending on which nuclides and whether known or suspected p-wave resonances are included in the analysis, in essence confirming results of our neutron-width analysis of the NDE. We also have examined radiation widths resulting from our Oak Ridge and Los Alamos measurements, and find that in some cases they do not agree with RMT. Although these disagreements presently are not understood, they could have broad impact on basic and applied nuclear physics, from nuclear astrophysics to nuclear criticality safety.Comment: 14 pages, 9 figures, submitted to special issue of Fortschritte Der Physik, Quantum Physics with Non-Hermitian Operator

    Equivalence of Systematic Linear Data Structures and Matrix Rigidity

    Get PDF
    Recently, Dvir, Golovnev, and Weinstein have shown that sufficiently strong lower bounds for linear data structures would imply new bounds for rigid matrices. However, their result utilizes an algorithm that requires an NPNP oracle, and hence, the rigid matrices are not explicit. In this work, we derive an equivalence between rigidity and the systematic linear model of data structures. For the nn-dimensional inner product problem with mm queries, we prove that lower bounds on the query time imply rigidity lower bounds for the query set itself. In particular, an explicit lower bound of ω(nrlogm)\omega\left(\frac{n}{r}\log m\right) for rr redundant storage bits would yield better rigidity parameters than the best bounds due to Alon, Panigrahy, and Yekhanin. We also prove a converse result, showing that rigid matrices directly correspond to hard query sets for the systematic linear model. As an application, we prove that the set of vectors obtained from rank one binary matrices is rigid with parameters matching the known results for explicit sets. This implies that the vector-matrix-vector problem requires query time Ω(n3/2/r)\Omega(n^{3/2}/r) for redundancy rnr \geq \sqrt{n} in the systematic linear model, improving a result of Chakraborty, Kamma, and Larsen. Finally, we prove a cell probe lower bound for the vector-matrix-vector problem in the high error regime, improving a result of Chattopadhyay, Kouck\'{y}, Loff, and Mukhopadhyay.Comment: 23 pages, 1 tabl

    Estimation of the Covariance Matrix of Large Dimensional Data

    Full text link
    This paper deals with the problem of estimating the covariance matrix of a series of independent multivariate observations, in the case where the dimension of each observation is of the same order as the number of observations. Although such a regime is of interest for many current statistical signal processing and wireless communication issues, traditional methods fail to produce consistent estimators and only recently results relying on large random matrix theory have been unveiled. In this paper, we develop the parametric framework proposed by Mestre, and consider a model where the covariance matrix to be estimated has a (known) finite number of eigenvalues, each of it with an unknown multiplicity. The main contributions of this work are essentially threefold with respect to existing results, and in particular to Mestre's work: To relax the (restrictive) separability assumption, to provide joint consistent estimates for the eigenvalues and their multiplicities, and to study the variance error by means of a Central Limit theorem
    corecore