480 research outputs found
Remarkable Effect of PEG-1000-based Dicationic Ionic Liquid for N-hydroxyphthalimide-catalyzed Aerobic Selective Oxidation of Alkylaromatics
PEG 1000–based functional dicationic acidic ionic liquid (PEG1000 –DAIL) was used for the first time as the reaction solvent for the N-Hydroxyphthalimide (NHPI)-cobalt acetate(Co(OAc)2) catalyzed aerobic oxidations of alkylaromatics to the corresponding acids. It enhanced the efficient catalytic ability of NHPI: 99.9 % conversion of toluene with 99.5 % selectivity for benzoic acid could be obtained at
80 °C in 10 h and ethylbenzene was selectively oxidized to benzoic acid. Several alkylaromatics were efficiently oxidized to their corresponding acids under mild conditions. For substituted toluene, the conversions of substrates and the selectivity of products was affected by the position and kind of substituted groups, respectively. Both the catalyst and PEG1000 –DAIL could be reused at least eight times without significantly decreasing the catalytic activity.(doi: 10.5562/cca2051
Who Should I Engage with At What Time? A Missing Event Aware Temporal Graph Neural Network
Temporal graph neural network has recently received significant attention due
to its wide application scenarios, such as bioinformatics, knowledge graphs,
and social networks. There are some temporal graph neural networks that achieve
remarkable results. However, these works focus on future event prediction and
are performed under the assumption that all historical events are observable.
In real-world applications, events are not always observable, and estimating
event time is as important as predicting future events. In this paper, we
propose MTGN, a missing event-aware temporal graph neural network, which
uniformly models evolving graph structure and timing of events to support
predicting what will happen in the future and when it will happen.MTGN models
the dynamic of both observed and missing events as two coupled temporal point
processes, thereby incorporating the effects of missing events into the
network. Experimental results on several real-world temporal graphs demonstrate
that MTGN significantly outperforms existing methods with up to 89% and 112%
more accurate time and link prediction. Code can be found on
https://github.com/HIT-ICES/TNNLS-MTGN.Comment: submitted to TNNL
A new graphical password scheme resistant to shoulder-surfing
Shoulder-surfing is a known risk where an attacker can capture a password by direct observation or by recording
the authentication session. Due to the visual interface, this problem has become exacerbated in graphical passwords.
There have been some graphical schemes resistant or immune
to shoulder-surfing, but they have significant usability
drawbacks, usually in the time and effort to log in. In this
paper, we propose and evaluate a new shoulder-surfing
resistant scheme which has a desirable usability for PDAs. Our inspiration comes from the drawing input method in DAS and the association mnemonics in Story for sequence retrieval. The new scheme requires users to draw a curve across their password images orderly rather than click directly on them.
The drawing input trick along with the complementary
measures, such as erasing the drawing trace, displaying
degraded images, and starting and ending with randomly
designated images provide a good resistance to shoulder-surfing.
A preliminary user study showed that users were able to enter their passwords accurately and to remember them over time
Can overfitted deep neural networks in adversarial training generalize? -- An approximation viewpoint
Adversarial training is a widely used method to improve the robustness of
deep neural networks (DNNs) over adversarial perturbations. However, it is
empirically observed that adversarial training on over-parameterized networks
often suffers from the \textit{robust overfitting}: it can achieve almost zero
adversarial training error while the robust generalization performance is not
promising. In this paper, we provide a theoretical understanding of the
question of whether overfitted DNNs in adversarial training can generalize from
an approximation viewpoint. Specifically, our main results are summarized into
three folds: i) For classification, we prove by construction the existence of
infinitely many adversarial training classifiers on over-parameterized DNNs
that obtain arbitrarily small adversarial training error (overfitting), whereas
achieving good robust generalization error under certain conditions concerning
the data quality, well separated, and perturbation level. ii) Linear
over-parameterization (meaning that the number of parameters is only slightly
larger than the sample size) is enough to ensure such existence if the target
function is smooth enough. iii) For regression, our results demonstrate that
there also exist infinitely many overfitted DNNs with linear
over-parameterization in adversarial training that can achieve almost optimal
rates of convergence for the standard generalization error. Overall, our
analysis points out that robust overfitting can be avoided but the required
model capacity will depend on the smoothness of the target function, while a
robust generalization gap is inevitable. We hope our analysis will give a
better understanding of the mathematical foundations of robustness in DNNs from
an approximation view
Research on Online Moisture Detector in Grain Drying Process Based on V/F Conversion
An online resistance grain moisture detector is designed, based on the model of the relationship between measurement frequency and grain moisture and the nonlinear correction method of temperature. The detector consists of lower computer, the core function of which is the sensing of grain resistance values which is based on V/F conversion, and upper computer, the core function of which is the conversion of moisture and frequency and the nonlinear correction of temperature. The performance of the online moisture detector is tested in a self-designed experimental system; the test and analysis results indicate that the precision and stability of the detector can reach the level of the similar products, which can be still improved
- …