578 research outputs found
Theory of optical imaging beyond the diffraction limit with a far-field superlens
Recent theoretical and experimental studies have shown that imaging with
resolution well beyond the diffraction limit can be obtained with so-called
superlenses. Images formed by such superlenses are, however, in the near field
only, or a fraction of wavelength away from the lens. In this paper, we propose
a far-field superlens (FSL) device which is composed of a planar superlens with
periodical corrugation. We show in theory that when an object is placed in
close proximity of such a FSL, a unique image can be formed in far-field. As an
example, we demonstrate numerically that images of 40 nm lines with a 30 nm gap
can be obtained from far-field data with properly designed FSL working at 376nm
wavelength.Comment: 6 pages, 3 figure
Direct photoluminescence probing of ferromagnetism in monolayer two-dimensional CrBr3
Atomically thin magnets are the key element to build up spintronics based on
two-dimensional materials. The surface nature of two-dimensional ferromagnet
opens up opportunities to improve the device performance efficiently. Here, we
report the intrinsic ferromagnetism in atomically thin monolayer CrBr3,
directly probed by polarization resolved magneto-photoluminescence. The
spontaneous magnetization persists in monolayer CrBr3 with a Curie temperature
of 34 K. The development of magnons by the thermal excitation is in line with
the spin-wave theory. We attribute the layer-number dependent hysteresis loops
in thick layers to the magnetic domain structures. As a stable monolayer
material in air, CrBr3 provides a convenient platform for fundamental physics
and pushes the potential applications of the two-dimensional ferromagnetism.Comment: 27 pages, 10 figure
Diet Code Is Healthy: Simplifying Programs for Pre-trained Models of Code
Pre-trained code representation models such as CodeBERT have demonstrated
superior performance in a variety of software engineering tasks, yet they are
often heavy in complexity, quadratically with the length of the input sequence.
Our empirical analysis of CodeBERT's attention reveals that CodeBERT pays more
attention to certain types of tokens and statements such as keywords and
data-relevant statements. Based on these findings, we propose DietCode, which
aims at lightweight leverage of large pre-trained models for source code.
DietCode simplifies the input program of CodeBERT with three strategies,
namely, word dropout, frequency filtering, and an attention-based strategy
which selects statements and tokens that receive the most attention weights
during pre-training. Hence, it gives a substantial reduction in the
computational cost without hampering the model performance. Experimental
results on two downstream tasks show that DietCodeBERT provides comparable
results to CodeBERT with 40% less computational cost in fine-tuning and
testing.Comment: Accepted to be published in ESEC/FSE 202
- …