873 research outputs found
原子スケール解析による CeO2立方体ナノ粒子の構造物性の解明
要約のみTohoku University阿尻雅文課
Depth Assisted Full Resolution Network for Single Image-based View Synthesis
Researches in novel viewpoint synthesis majorly focus on interpolation from
multi-view input images. In this paper, we focus on a more challenging and
ill-posed problem that is to synthesize novel viewpoints from one single input
image. To achieve this goal, we propose a novel deep learning-based technique.
We design a full resolution network that extracts local image features with the
same resolution of the input, which contributes to derive high resolution and
prevent blurry artifacts in the final synthesized images. We also involve a
pre-trained depth estimation network into our system, and thus 3D information
is able to be utilized to infer the flow field between the input and the target
image. Since the depth network is trained by depth order information between
arbitrary pairs of points in the scene, global image features are also involved
into our system. Finally, a synthesis layer is used to not only warp the
observed pixels to the desired positions but also hallucinate the missing
pixels with recorded pixels. Experiments show that our technique performs well
on images of various scenes, and outperforms the state-of-the-art techniques
Language Models for Image Captioning: The Quirks and What Works
Two recent approaches have achieved state-of-the-art results in image
captioning. The first uses a pipelined process where a set of candidate words
is generated by a convolutional neural network (CNN) trained on images, and
then a maximum entropy (ME) language model is used to arrange these words into
a coherent sentence. The second uses the penultimate activation layer of the
CNN as input to a recurrent neural network (RNN) that then generates the
caption sequence. In this paper, we compare the merits of these different
language modeling approaches for the first time by using the same
state-of-the-art CNN as input. We examine issues in the different approaches,
including linguistic irregularities, caption repetition, and data set overlap.
By combining key aspects of the ME and RNN methods, we achieve a new record
performance over previously published results on the benchmark COCO dataset.
However, the gains we see in BLEU do not translate to human judgments.Comment: See http://research.microsoft.com/en-us/projects/image_captioning for
project informatio
Automatic Hallucination Assessment for Aligned Large Language Models via Transferable Adversarial Attacks
Although remarkable progress has been achieved in preventing large language
model (LLM) hallucinations using instruction tuning and retrieval augmentation,
it remains challenging to measure the reliability of LLMs using human-crafted
evaluation data which is not available for many tasks and domains and could
suffer from data leakage. Inspired by adversarial machine learning, this paper
aims to develop a method of automatically generating evaluation data by
appropriately modifying existing data on which LLMs behave faithfully.
Specifically, this paper presents AutoDebug, an LLM-based framework to use
prompting chaining to generate transferable adversarial attacks in the form of
question-answering examples. We seek to understand the extent to which these
examples trigger the hallucination behaviors of LLMs.
We implement AutoDebug using ChatGPT and evaluate the resulting two variants
of a popular open-domain question-answering dataset, Natural Questions (NQ), on
a collection of open-source and proprietary LLMs under various prompting
settings. Our generated evaluation data is human-readable and, as we show,
humans can answer these modified questions well. Nevertheless, we observe
pronounced accuracy drops across multiple LLMs including GPT-4. Our
experimental results show that LLMs are likely to hallucinate in two categories
of question-answering scenarios where (1) there are conflicts between knowledge
given in the prompt and their parametric knowledge, or (2) the knowledge
expressed in the prompt is complex. Finally, we find that the adversarial
examples generated by our method are transferable across all considered LLMs.
The examples generated by a small model can be used to debug a much larger
model, making our approach cost-effective
Metabolome and transcriptome analyses reveal the colouring mechanism of red honeysuckle (Lonicera japonica Thunb.)
Honeysuckle has been widely used as a medicinal herb and food additive in China for a long time. However, little is known about the pigment composition and colouring mechanism of red honeysuckle, which is a rare germplasm resource. This study aims to investigate the anthocyanin components and colouring mechanism of red honeysuckle, and to identify potential regulatory genes in the anthocyanin biosynthesis pathway. ‘Yujin 1’ and ‘Yujin 2’, with yellow-white and red flower buds, respectively, were selected for the study. Using a metabolomics method, we identified the anthocyanin components, while transcriptomics analysis was used to mine the structural and regulatory genes of the anthocyanin biosynthesis pathway. Additionally, protein-protein interaction analysis was employed to predict the regulation mechanism of anthocyanin biosynthesis. The results revealed that cyanidin-3,5-O-diglucoside, peonidin-3,5-O-diglucoside, and cyanidin-3-O-glucoside were the main pigment components of red honeysuckle. We also constructed a possible anthocyanin biosynthetic pathway and identified MYB and bHLH transcription factors that may play regulatory roles in this pathway. Furthermore, our findings suggest that bHLH23 may regulate anthocyanin biosynthesis by binding to the DFR gene promoter. These findings have significant implications for breeding new honeysuckle varieties and developing functional foods and medicines
A Novel Ehanced Move Recognition Algorithm Based on Pre-trained Models with Positional Embeddings
The recognition of abstracts is crucial for effectively locating the content
and clarifying the article. Existing move recognition algorithms lack the
ability to learn word position information to obtain contextual semantics. This
paper proposes a novel enhanced move recognition algorithm with an improved
pre-trained model and a gated network with attention mechanism for unstructured
abstracts of Chinese scientific and technological papers. The proposed
algorithm first performs summary data segmentation and vocabulary training. The
EP-ERNIEAT-GRU framework is leveraged to incorporate word positional
information, facilitating deep semantic learning and targeted feature
extraction. Experimental results demonstrate that the proposed algorithm
achieves 13.37 higher accuracy on the split dataset than on the original
dataset and a 7.55 improvement in accuracy over the basic comparison model
Soil Liquid Limit and Plastic Limit Treating System Based on Analytic Method
AbstractAccording to two present China national standards, a software as Soil Liquid Limit and Plastic Limit Data Treating System, with analytic method, was developed using object-oriented visual programming tool. The analytic method used in the developed system was different to traditional method of treating soil liquid limit and plastic limit data. N-S algorithm flowchart demonstrated that switch statement and condition statement were taken as main algorithm and second level select nested structure was taken as main frame for the developed system. Three kinds of soil specimens were tested with liquid and plastic limit combined test and the test data was treated with graphic method, Excel software and Soil Liquid Limit and Plastic Limit Data Treating System. The comparative conclusion indicated that Soil Liquid Limit and Plastic Limit Data Treating System improved efficiency and accuracy evidently for treating soil liquid and plastic limit data and had advantages of easy operation and high reliability
Learning for Semantic Knowledge Base-Guided Online Feature Transmission in Dynamic Channels
With the proliferation of edge computing, efficient AI inference on edge
devices has become essential for intelligent applications such as autonomous
vehicles and VR/AR. In this context, we address the problem of efficient remote
object recognition by optimizing feature transmission between mobile devices
and edge servers. We propose an online optimization framework to address the
challenge of dynamic channel conditions and device mobility in an end-to-end
communication system. Our approach builds upon existing methods by leveraging a
semantic knowledge base to drive multi-level feature transmission, accounting
for temporal factors and dynamic elements throughout the transmission process.
To solve the online optimization problem, we design a novel soft
actor-critic-based deep reinforcement learning system with a carefully designed
reward function for real-time decision-making, overcoming the optimization
difficulty of the NP-hard problem and achieving the minimization of semantic
loss while respecting latency constraints. Numerical results showcase the
superiority of our approach compared to traditional greedy methods under
various system setups.Comment: 6 page
- …