1,122 research outputs found
Octahedral Chiral-at-Metal Iridium and Rhodium Complexes as Versatile Asymmetric Catalysts
Over the past several years, our group has been interested in designing and synthesizing different novel octahedral chiral-at-metal complexes and their application in asymmetric catalysis, including visible-light-induced asymmetric catalysis.
This thesis mainly includes two parts: one is the versatile asymmetric catalysis by octahedral chiral-at-metal iridium complexes, and the other one is the visible-light-promoted asymmetric α-amination by octahedral chiral-at-metal rhodium complex.
In the first part of this thesis, octahedral chiral-at-metal iridium complexes IrS and IrO are used as highly effective chiral Lewis acid catalysts for a variety of asymmetric reactions, including Friedel-Crafts alkylations, Michael additions with CH-acidic compounds, 1,3-dipolar cycloadditions, Diels Alder cycloadditions, hetero Diels Alder cycloadditions and Henry reactions.
In the second part of the thesis, a very efficient photoactivated enantioselective radical amination of 2-acyl imidazoles catalyzed by an octahedral chiral-at-metal rhodium complex RhO is introduced. Rhodium complex here serves a dual function, namely as a chiral Lewis acid to catalyze asymmetric enolate chemistry and furthermore as a light-activated smart initiator of a radical chain process
Stochastic Answer Networks for Machine Reading Comprehension
We propose a simple yet robust stochastic answer network (SAN) that simulates
multi-step reasoning in machine reading comprehension. Compared to previous
work such as ReasoNet which used reinforcement learning to determine the number
of steps, the unique feature is the use of a kind of stochastic prediction
dropout on the answer module (final layer) of the neural network during the
training. We show that this simple trick improves robustness and achieves
results competitive to the state-of-the-art on the Stanford Question Answering
Dataset (SQuAD), the Adversarial SQuAD, and the Microsoft MAchine Reading
COmprehension Dataset (MS MARCO).Comment: 11 pages, 5 figures, Accepted to ACL 201
Language-Based Image Editing with Recurrent Attentive Models
We investigate the problem of Language-Based Image Editing (LBIE). Given a
source image and a natural language description, we want to generate a target
image by editing the source image based on the description. We propose a
generic modeling framework for two sub-tasks of LBIE: language-based image
segmentation and image colorization. The framework uses recurrent attentive
models to fuse image and language features. Instead of using a fixed step size,
we introduce for each region of the image a termination gate to dynamically
determine after each inference step whether to continue extrapolating
additional information from the textual description. The effectiveness of the
framework is validated on three datasets. First, we introduce a synthetic
dataset, called CoSaL, to evaluate the end-to-end performance of our LBIE
system. Second, we show that the framework leads to state-of-the-art
performance on image segmentation on the ReferIt dataset. Third, we present the
first language-based colorization result on the Oxford-102 Flowers dataset.Comment: Accepted to CVPR 2018 as a Spotligh
A Deep Embedding Model for Co-occurrence Learning
Co-occurrence Data is a common and important information source in many
areas, such as the word co-occurrence in the sentences, friends co-occurrence
in social networks and products co-occurrence in commercial transaction data,
etc, which contains rich correlation and clustering information about the
items. In this paper, we study co-occurrence data using a general energy-based
probabilistic model, and we analyze three different categories of energy-based
model, namely, the , and models, which are able to capture
different levels of dependency in the co-occurrence data. We also discuss how
several typical existing models are related to these three types of energy
models, including the Fully Visible Boltzmann Machine (FVBM) (), Matrix
Factorization (), Log-BiLinear (LBL) models (), and the Restricted
Boltzmann Machine (RBM) model (). Then, we propose a Deep Embedding Model
(DEM) (an model) from the energy model in a \emph{principled} manner.
Furthermore, motivated by the observation that the partition function in the
energy model is intractable and the fact that the major objective of modeling
the co-occurrence data is to predict using the conditional probability, we
apply the \emph{maximum pseudo-likelihood} method to learn DEM. In consequence,
the developed model and its learning method naturally avoid the above
difficulties and can be easily used to compute the conditional probability in
prediction. Interestingly, our method is equivalent to learning a special
structured deep neural network using back-propagation and a special sampling
strategy, which makes it scalable on large-scale datasets. Finally, in the
experiments, we show that the DEM can achieve comparable or better results than
state-of-the-art methods on datasets across several application domains
MC-Nonlocal-PINNs: handling nonlocal operators in PINNs via Monte Carlo sampling
We propose, Monte Carlo Nonlocal physics-informed neural networks
(MC-Nonlocal-PINNs), which is a generalization of MC-fPINNs in
\cite{guo2022monte}, for solving general nonlocal models such as integral
equations and nonlocal PDEs. Similar as in MC-fPINNs, our MC-Nonlocal-PINNs
handle the nonlocal operators in a Monte Carlo way, resulting in a very stable
approach for high dimensional problems. We present a variety of test problems,
including high dimensional Volterra type integral equations, hypersingular
integral equations and nonlocal PDEs, to demonstrate the effectiveness of our
approach.Comment: 23pages, 13figure
Diet Code Is Healthy: Simplifying Programs for Pre-trained Models of Code
Pre-trained code representation models such as CodeBERT have demonstrated
superior performance in a variety of software engineering tasks, yet they are
often heavy in complexity, quadratically with the length of the input sequence.
Our empirical analysis of CodeBERT's attention reveals that CodeBERT pays more
attention to certain types of tokens and statements such as keywords and
data-relevant statements. Based on these findings, we propose DietCode, which
aims at lightweight leverage of large pre-trained models for source code.
DietCode simplifies the input program of CodeBERT with three strategies,
namely, word dropout, frequency filtering, and an attention-based strategy
which selects statements and tokens that receive the most attention weights
during pre-training. Hence, it gives a substantial reduction in the
computational cost without hampering the model performance. Experimental
results on two downstream tasks show that DietCodeBERT provides comparable
results to CodeBERT with 40% less computational cost in fine-tuning and
testing.Comment: Accepted to be published in ESEC/FSE 202
- …