21,458 research outputs found
Relaxed 2-D Principal Component Analysis by Norm for Face Recognition
A relaxed two dimensional principal component analysis (R2DPCA) approach is
proposed for face recognition. Different to the 2DPCA, 2DPCA- and G2DPCA,
the R2DPCA utilizes the label information (if known) of training samples to
calculate a relaxation vector and presents a weight to each subset of training
data. A new relaxed scatter matrix is defined and the computed projection axes
are able to increase the accuracy of face recognition. The optimal -norms
are selected in a reasonable range. Numerical experiments on practical face
databased indicate that the R2DPCA has high generalization ability and can
achieve a higher recognition rate than state-of-the-art methods.Comment: 19 pages, 11 figure
On the Convergence of Ritz Pairs and Refined Ritz Vectors for Quadratic Eigenvalue Problems
For a given subspace, the Rayleigh-Ritz method projects the large quadratic
eigenvalue problem (QEP) onto it and produces a small sized dense QEP. Similar
to the Rayleigh-Ritz method for the linear eigenvalue problem, the
Rayleigh-Ritz method defines the Ritz values and the Ritz vectors of the QEP
with respect to the projection subspace. We analyze the convergence of the
method when the angle between the subspace and the desired eigenvector
converges to zero. We prove that there is a Ritz value that converges to the
desired eigenvalue unconditionally but the Ritz vector converges conditionally
and may fail to converge. To remedy the drawback of possible non-convergence of
the Ritz vector, we propose a refined Ritz vector that is mathematically
different from the Ritz vector and is proved to converge unconditionally. We
construct examples to illustrate our theory.Comment: 20 page
A general software defect-proneness prediction framework
This is the author's accepted manuscript. The final published article is available from the link below. Copyright @ 2011 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works.BACKGROUND - Predicting defect-prone software components is an economically important activity and so has received a good deal of attention. However, making sense of the many, and sometimes seemingly inconsistent, results is difficult. OBJECTIVE - We propose and evaluate a general framework for software defect prediction that supports 1) unbiased and 2) comprehensive comparison between competing prediction systems. METHOD - The framework is comprised of 1) scheme evaluation and 2) defect prediction components. The scheme evaluation analyzes the prediction performance of competing learning schemes for given historical data sets. The defect predictor builds models according to the evaluated learning scheme and predicts software defects with new data according to the constructed model. In order to demonstrate the performance of the proposed framework, we use both simulation and publicly available software defect data sets. RESULTS - The results show that we should choose different learning schemes for different data sets (i.e., no scheme dominates), that small details in conducting how evaluations are conducted can completely reverse findings, and last, that our proposed framework is more effective and less prone to bias than previous approaches. CONCLUSIONS - Failure to properly or fully evaluate a learning scheme can be misleading; however, these problems may be overcome by our proposed framework.National Natural Science Foundation of
Chin
Vacuum induced Berry phases in single-mode Jaynes-Cummings models
Motivated by the work [Phys. Rev. Lett. 89, 220404 (2002)] for detecting the
vacuum-induced Berry phases with two-mode Jaynes-Cummings models (JCMs), we
show here that, for a parameter-dependent single-mode JCM, certain atom-field
states also acquire the photon-number-dependent Berry phases after the
parameter slowly changed and eventually returned to its initial value. This
geometric effect related to the field quantization still exists, even the filed
is kept in its vacuum state. Specifically, a feasible Ramsey interference
experiment with cavity quantum electrodynamics (QED) system is designed to
detect the vacuum-induced Berry phase.Comment: 10 pages, 4 figures
TEQUILA: Temporal Question Answering over Knowledge Bases
Question answering over knowledge bases (KB-QA) poses challenges in handling complex questions that need to be decomposed into sub-questions. An important case, addressed here, is that of temporal questions, where cues for temporal relations need to be discovered and handled. We present TEQUILA, an enabler method for temporal QA that can run on top of any KB-QA engine. TEQUILA has four stages. It detects if a question has temporal intent. It decomposes and rewrites the question into non-temporal sub-questions and temporal constraints. Answers to sub-questions are then retrieved from the underlying KB-QA engine. Finally, TEQUILA uses constraint reasoning on temporal intervals to compute final answers to the full question. Comparisons against state-of-the-art baselines show the viability of our method
BDGS: A Scalable Big Data Generator Suite in Big Data Benchmarking
Data generation is a key issue in big data benchmarking that aims to generate
application-specific data sets to meet the 4V requirements of big data.
Specifically, big data generators need to generate scalable data (Volume) of
different types (Variety) under controllable generation rates (Velocity) while
keeping the important characteristics of raw data (Veracity). This gives rise
to various new challenges about how we design generators efficiently and
successfully. To date, most existing techniques can only generate limited types
of data and support specific big data systems such as Hadoop. Hence we develop
a tool, called Big Data Generator Suite (BDGS), to efficiently generate
scalable big data while employing data models derived from real data to
preserve data veracity. The effectiveness of BDGS is demonstrated by developing
six data generators covering three representative data types (structured,
semi-structured and unstructured) and three data sources (text, graph, and
table data)
BRDFs acquired by directional radiative measurements during EAGLE and AGRISAR
Radiation is the driving force for all processes and interactions between earth surface and atmosphere. The amount of
measured radiation reflected by vegetation depends on its structure, the viewing angle and the solar angle. This angular
dependence is usually expressed in the Bi-directional Reflectance Distribution Function (BRDF). This BRDF is not
only different for different types of vegetation, but also different for different stages of the growth. The BRDF therefore
has to be measured at ground level before any satellite imagery can be used the calculate surface-atmosphere
interaction. The objective of this research is to acquire the BRDFs for agricultural crop types.
A goniometric system is used to acquire the BRDFs. This is a mechanical device capable of a complete hemispherical
rotation. The radiative directional measurements are performed with different sensors that can be attached to this
system. The BRDFs are calculated from the measured radiation.
In the periods 10 June - 18 June 2006 and 2 July - 10 July 2006 directional radiative measurements were performed at
three sites: Speulderbos site, in the Netherlands, the Cabauw site, in the Netherlands, and an agricultural test site in
Goermin, Germany. The measurements were performed over eight different crops: forest, grass, pine tree, corn, wheat,
sugar beat and barley. The sensors covered the spectrum from the optical to the thermal domain. The measured radiance
is used to calculate the BRDFs or directional thermal signature.
This contribution describes the measurements and calculation of the BRDFs of forest, grassland, young corn, mature
corn, wheat, sugar beat and barley during the EAGLE2006 and AGRISAR 2006 fieldcampaigns. Optical BRDF have
been acquired for all crops except barley. Thermal angular signatures are acquired for all the crop
Dividend Payments and Excess Cash: An Experimental Analysis
There is an observed positive correlation between the size of firms’ dividends and the amount of cash they keep on hand. We specify a new channel of precautionary cash holdings as related to dividend-smoothing, which predicts that increased dividends should cause firms to keep more cash on hand. In the reverse direction, theory’s predictions depend on the source of the increased cash on hand. General dividend smoothing theory predicts that temporary increases in cash due to natural market fluctuations should not affect dividends. If the increased cash is due to a permanent increase in profitability, both the marginal benefit and the marginal cost of paying dividends also increase, and the net effect is unclear. Theory’s predictions of causality cannot be tested empirically using observed firm data, however, due to an underlying endogeneity issue. We instead turn to a laboratory experiment to determine causality. In the lab, we separately and exogenously vary each relevant variable to test each theoretical hypothesis. Increased dividends lead to increased cash on hand. One-time increases in firms’ cash on hand, representing natural market fluctuations, do not affect firms’ dividends. Increases in cash on hand due to a permanent increase in firm profitability, however, cause dividends to decrease
Being Watched in an Investment Game Setting: Behavioral Changes when Making Risky Decisions
We design a laboratory experiment to test for behavioral differences due to observation within a novel arena: investment games. We find that fund managers are more risk-averse when investors can observe their investment allocations. This effect is more pronounced when investors, in addition to observing the allocations, can also observe the investment outcomes. Interestingly, allowing investors to observe how their investment is allocated does not impact how much they invest. Last, when the outcome of the risky investment is public knowledge, disclosing managers’ allocations leads them to return more tokens to investors and to expropriate fewer tokens for themselves at the end of the game, ceteris paribus. We discuss potential causes of these effects
- …