187 research outputs found
Study on Optimized Elman Neural Network Classification Algorithm Based on PLS and CA
High-dimensional large sample data sets, between feature variables and between samples, may cause some correlative or repetitive factors, occupy lots of storage space, and consume much computing time. Using the Elman neural network to deal with them, too many inputs will influence the operating efficiency and recognition accuracy; too many simultaneous training samples, as well as being not able to get precise neural network model, also restrict the recognition accuracy. Aiming at these series of problems, we introduce the partial least squares (PLS) and cluster analysis (CA) into Elman neural network algorithm, by the PLS for dimension reduction which can eliminate the correlative and repetitive factors of the features. Using CA eliminates the correlative and repetitive factors of the sample. If some subclass becomes small sample, with high-dimensional feature and fewer numbers, PLS shows a unique advantage. Each subclass is regarded as one training sample to train the different precise neural network models. Then simulation samples are discriminated and classified into different subclasses, using the corresponding neural network to recognize it. An optimized Elman neural network classification algorithm based on PLS and CA (PLS-CA-Elman algorithm) is established. The new algorithm aims at improving the operating efficiency and recognition accuracy. By the case analysis, the new algorithm has unique superiority, worthy of further promotion
Detoxify Language Model Step-by-Step
Detoxification for LLMs is challenging since it requires models to avoid
generating harmful content while maintaining the generation capability. To
ensure the safety of generations, previous detoxification methods detoxify the
models by changing the data distributions or constraining the generations from
different aspects in a single-step manner. However, these approaches will
dramatically affect the generation quality of LLMs, e.g., discourse coherence
and semantic consistency, since language models tend to generate along the
toxic prompt while detoxification methods work in the opposite direction. To
handle such a conflict, we decompose the detoxification process into different
sub-steps, where the detoxification is concentrated in the input stage and the
subsequent continual generation is based on the non-toxic prompt. Besides, we
also calibrate the strong reasoning ability of LLMs by designing a Detox-Chain
to connect the above sub-steps in an orderly manner, which allows LLMs to
detoxify the text step-by-step. Automatic and human evaluation on two
benchmarks reveals that by training with Detox-Chain, six LLMs scaling from 1B
to 33B can obtain significant detoxification and generation improvement. Our
code and data are available at https://github.com/CODINNLG/Detox-CoT. Warning:
examples in the paper may contain uncensored offensive content
Wearable and Invisible Sensor Design for Eye-Motion Monitoring Based on Ferrofluid and Electromagnetic Sensing Technologies
For many human body diseases, treatments in the early stages are more efficient and safer than those in the later stages; therefore, detecting the early symptoms of a disease is crucial. One of the most significant early indicators for diseases is bio-mechanical motion. This paper provides a unique way of monitoring bio-mechanical eye motion based on electromagnetic sensing technology and a ferro-magnetic material, ferrofluid. The proposed monitoring method has the advantages of being inexpensive, non-invasive, sensor-invisible and extremely effective. Most of the medical devices are cumbersome and bulky, which makes them hard to apply for daily monitoring. However, the proposed eye-motion monitoring method is designed based on ferrofluid eye make-up and invisible sensors embedded inside the frame of glasses such that the system is wearable for daily monitoring. In addition, it has no influence on the appearance of the patient, which is beneficial for the mental health of some patients who do not want to attract public attention during treatment. The sensor responses are modelled using finite element simulation models, and wearable sensor systems are created. The designed frame of the glasses is manufactured based on 3-D printing technology. Experiments are conducted to monitor eye bio-mechanical motions, such as the frequency of eye blinking. Both the quick blinking behaviour with an overall frequency of around 1.1 Hz and the slow blinking behaviour with an overall frequency of around 0.4 Hz can be observed through experimentation. Simulations and measurements results show that the proposed sensor design can be employed for bio-mechanical eye-motion monitoring. In addition, the proposed system has the advantages of invisible sensor set-up and will not affect the appearance of the patient, which is not only convenient for the daily life of the patient but also beneficial for mental health
ScalAna: Automating Scaling Loss Detection with Graph Analysis
Scaling a parallel program to modern supercomputers is challenging due to
inter-process communication, Amdahl's law, and resource contention. Performance
analysis tools for finding such scaling bottlenecks either base on profiling or
tracing. Profiling incurs low overheads but does not capture detailed
dependencies needed for root-cause analysis. Tracing collects all information
at prohibitive overheads. In this work, we design ScalAna that uses static
analysis techniques to achieve the best of both worlds - it enables the
analyzability of traces at a cost similar to profiling. ScalAna first leverages
static compiler techniques to build a Program Structure Graph, which records
the main computation and communication patterns as well as the program's
control structures. At runtime, we adopt lightweight techniques to collect
performance data according to the graph structure and generate a Program
Performance Graph. With this graph, we propose a novel approach, called
backtracking root cause detection, which can automatically and efficiently
detect the root cause of scaling loss. We evaluate ScalAna with real
applications. Results show that our approach can effectively locate the root
cause of scaling loss for real applications and incurs 1.73% overhead on
average for up to 2,048 processes. We achieve up to 11.11% performance
improvement by fixing the root causes detected by ScalAna on 2,048 processes.Comment: conferenc
HIV Exploits Antiviral Host Innate GCN2-ATF4 Signaling for Establishing Viral Replication Early in Infection.
Antiviral innate host defenses against acute viral infections include suppression of host protein synthesis to restrict viral protein production. Less is known about mechanisms by which viral pathogens subvert host antiviral innate responses for establishing their replication and dissemination. We investigated early innate defense against human immunodeficiency virus (HIV) infection and viral evasion by utilizing human CD4+ T cell cultures in vitro and a simian immunodeficiency virus (SIV) model of AIDS in vivo Our data showed that early host innate defense against the viral infection involves GCN2-ATF4 signaling-mediated suppression of global protein synthesis, which is exploited by the virus for supporting its own replication during early viral infection and dissemination in the gut mucosa. Suppression of protein synthesis and induction of protein kinase GCN2-ATF4 signaling were detected in the gut during acute SIV infection. These changes diminished during chronic viral infection. HIV replication induced by serum deprivation in CD4+ T cells was linked to the induction of ATF4 that was recruited to the HIV long terminal repeat (LTR) to promote viral transcription. Experimental inhibition of GCN2-ATF4 signaling either by a specific inhibitor or by amino acid supplementation suppressed the induction of HIV expression. Enhancing ATF4 expression through selenium administration resulted in reactivation of latent HIV in vitro as well as ex vivo in the primary CD4+ T cells isolated from patients receiving suppressive antiretroviral therapy (ART). In summary, HIV/SIV exploits the early host antiviral response through GCN2-ATF4 signaling by utilizing ATF4 for activating the viral LTR transcription to establish initial viral replication and is a potential target for HIV prevention and therapy.IMPORTANCE Understanding how HIV overcomes host antiviral innate defense response in order to establish infection and dissemination is critical for developing prevention and treatment strategies. Most investigations focused on the viral pathogenic mechanisms leading to immune dysfunction following robust viral infection and dissemination. Less is known about mechanisms that enable HIV to establish its presence despite rapid onset of host antiviral innate response. Our novel findings provide insights into the viral strategy that hijacks the host innate response of the suppression of protein biosynthesis to restrict the virus production. The virus leverages transcription factor ATF4 expression during the GCN2-ATF4 signaling response and utilizes it to activate viral transcription through the LTR to support viral transcription and production in both HIV and SIV infections. This unique viral strategy is exploiting the innate response and is distinct from the mechanisms of immune dysfunction after the critical mass of viral loads is generated
Towards Lightweight and Automated Representation Learning System for Networks
We propose LIGHTNE 2.0, a cost-effective, scalable, automated, and
high-quality network embedding system that scales to graphs with hundreds of
billions of edges on a single machine. In contrast to the mainstream belief
that distributed architecture and GPUs are needed for large-scale network
embedding with good quality, we prove that we can achieve higher quality,
better scalability, lower cost, and faster runtime with shared-memory, CPU-only
architecture. LIGHTNE 2.0 combines two theoretically grounded embedding methods
NetSMF and ProNE. We introduce the following techniques to network embedding
for the first time: (1) a newly proposed downsampling method to reduce the
sample complexity of NetSMF while preserving its theoretical advantages; (2) a
high-performance parallel graph processing stack GBBS to achieve high memory
efficiency and scalability; (3) sparse parallel hash table to aggregate and
maintain the matrix sparsifier in memory; (4) a fast randomized singular value
decomposition (SVD) enhanced by power iteration and fast orthonormalization to
improve vanilla randomized SVD in terms of both efficiency and effectiveness;
(5) Intel MKL for proposed fast randomized SVD and spectral propagation; and
(6) a fast and lightweight AutoML library FLAML for automated hyperparameter
tuning. Experimental results show that LIGHTNE 2.0 can be up to 84X faster than
GraphVite, 30X faster than PBG and 9X faster than NetSMF while delivering
better performance. LIGHTNE 2.0 can embed very large graph with 1.7 billion
nodes and 124 billion edges in half an hour on a CPU server, while other
baselines cannot handle very large graphs of this scale
Defining a framework for the evaluation of information
In any enterprise, principled decisions need be made during the entire life
cycle of information about its acquisition, storage, creation, maintenance and disposal.
Such information management requires some form of information evaluation to take
place, yet little is understood about the process of information evaluation within
enterprises. For evaluation support to be both effective and resource efficient,
particularly where decisions are being made about the future of large quantities of
information, it would be invaluable if some sort of automatic or semi-automatic
methods were available for evaluation. Such a method would require an understanding
of the diversity of the contexts in which evaluation takes place so that evaluation
support can have the necessary context-sensitivity. This paper identifies the dimensions
that influence the information evaluation process and defines the elements that
characterize these dimensions, thus providing the foundations for a context-sensitive
framework for information evaluation
- …