72,206 research outputs found
Josephson effect in spin-singlet superconductor/ferromagnetic insulator/spin-triplet superconductor junctions with helical -wave states
We study the Josephson effect in spin-singlet superconductor/helical -wave
superconductor junctions with a ferromagnetic barrier using the quasiclassical
Green function method. It is found that both -type and
-type current-phase relations always exist, irrespective of the gap
symmetries in superconductors. The indispensable condition for the
-type and -type current is that the magnetization must
have a component parallel to the crystallographic or axis, which is
distinct from the case of -wave superconductor described by a
\vect{d}-vector with a uniform direction. The relation between the condition
and the symmetries of the gap functions is analysed. We investigate in detail
the symmetries and the sign reversal of the Josephson current when the
magnetization is rotated.Comment: 9 pages,7 figure
A Generic Position Based Method for Real Root Isolation of Zero-Dimensional Polynomial Systems
We improve the local generic position method for isolating the real roots of
a zero-dimensional bivariate polynomial system with two polynomials and extend
the method to general zero-dimensional polynomial systems. The method mainly
involves resultant computation and real root isolation of univariate polynomial
equations. The roots of the system have a linear univariate representation. The
complexity of the method is for the bivariate case, where
, resp., is an upper bound on the degree, resp., the
maximal coefficient bitsize of the input polynomials. The algorithm is
certified with probability 1 in the multivariate case. The implementation shows
that the method is efficient, especially for bivariate polynomial systems.Comment: 24 pages, 5 figure
Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)
Recently, increasing works have proposed to drive evolutionary algorithms
using machine learning models. Usually, the performance of such model based
evolutionary algorithms is highly dependent on the training qualities of the
adopted models. Since it usually requires a certain amount of data (i.e. the
candidate solutions generated by the algorithms) for model training, the
performance deteriorates rapidly with the increase of the problem scales, due
to the curse of dimensionality. To address this issue, we propose a
multi-objective evolutionary algorithm driven by the generative adversarial
networks (GANs). At each generation of the proposed algorithm, the parent
solutions are first classified into real and fake samples to train the GANs;
then the offspring solutions are sampled by the trained GANs. Thanks to the
powerful generative ability of the GANs, our proposed algorithm is capable of
generating promising offspring solutions in high-dimensional decision space
with limited training data. The proposed algorithm is tested on 10 benchmark
problems with up to 200 decision variables. Experimental results on these test
problems demonstrate the effectiveness of the proposed algorithm
Stacking-based Deep Neural Network: Deep Analytic Network on Convolutional Spectral Histogram Features
Stacking-based deep neural network (S-DNN), in general, denotes a deep neural
network (DNN) resemblance in terms of its very deep, feedforward network
architecture. The typical S-DNN aggregates a variable number of individually
learnable modules in series to assemble a DNN-alike alternative to the targeted
object recognition tasks. This work likewise devises an S-DNN instantiation,
dubbed deep analytic network (DAN), on top of the spectral histogram (SH)
features. The DAN learning principle relies on ridge regression, and some key
DNN constituents, specifically, rectified linear unit, fine-tuning, and
normalization. The DAN aptitude is scrutinized on three repositories of varying
domains, including FERET (faces), MNIST (handwritten digits), and CIFAR10
(natural objects). The empirical results unveil that DAN escalates the SH
baseline performance over a sufficiently deep layer.Comment: 5 page
- …
