71 research outputs found

    Measurement of permeability for ferrous metallic plates using a novel lift-off compensation technique on phase signature

    Full text link
    Lift-off of sensor affects the prediction of electromagnetic properties for both ferrous and non-ferrous steel plates. In this paper, we developed a strategy to address this issue for ferrous plates. With increased lift-off, the phase of the measured impedance for steel plates reduces. Meanwhile, the magnitude of the impedance signal decreases. Based on these facts, a phase compensation algorithm is developed which corrects the phase change due to lift-off considering the magnitude of the impedance signal. Further, a new magnetic permeability prediction technique is presented, which has been validated by analytical and measured results. With this new technique, the error in permeability prediction is less than 2% within the range of lift-offs tested

    Learning to Initialize: Can Meta Learning Improve Cross-task Generalization in Prompt Tuning?

    Full text link
    Prompt tuning (PT) which only tunes the embeddings of an additional sequence of tokens per task, keeping the pre-trained language model (PLM) frozen, has shown remarkable performance in few-shot learning. Despite this, PT has been shown to rely heavily on good initialization of the prompt embeddings. In this work, we study meta prompt tuning (MPT) to systematically explore how meta-learning can help improve (if it can) cross-task generalization in PT through learning to initialize the prompt embeddings from other relevant tasks. We empirically analyze a representative set of meta learning algorithms in a wide range of adaptation settings with different source/target task configurations on a large set of few-shot tasks. With extensive experiments and analysis, we demonstrate the effectiveness of MPT. We find the improvement to be significant particularly on classification tasks. For other kinds of tasks such as question answering, we observe that while MPT can outperform PT in most cases, it does not always outperform multi-task learning. We further provide an in-depth analysis from the perspective of task similarity

    An equivalent-effect phenomenon in eddy current non-destructive testing of thin structures

    Full text link
    The inductance/impedance due to thin metallic structures in non-destructive testing (NDT) is difficult to evaluate. In particular, in Finite Element Method (FEM) eddy current simulation, an extremely fine mesh is required to accurately simulate skin effects especially at high frequencies, and this could cause an extremely large total mesh for the whole problem, i.e. including, for example, other surrounding structures and excitation sources like coils. Consequently, intensive computation requirements are needed. In this paper, an equivalent-effect phenomenon is found, which has revealed that alternative structures can produce the same effect on the sensor response, i.e. mutual impedance/inductance of coupled coils if a relationship (reciprocal relationship) between the electrical conductivity and the thickness of the structure is observed. By using this relationship, the mutual inductance/impedance can be calculated from the equivalent structures with much fewer mesh elements, which can significantly save the computation time. In eddy current NDT, coils inductance/impedance is normally used as a critical parameter for various industrial applications, such as flaw detection, coating and microstructure sensing. Theoretical derivation, measurements and simulations have been presented to verify the feasibility of the proposed phenomenon

    Lifelong Event Detection with Embedding Space Separation and Compaction

    Full text link
    To mitigate forgetting, existing lifelong event detection methods typically maintain a memory module and replay the stored memory data during the learning of a new task. However, the simple combination of memory data and new-task samples can still result in substantial forgetting of previously acquired knowledge, which may occur due to the potential overlap between the feature distribution of new data and the previously learned embedding space. Moreover, the model suffers from overfitting on the few memory samples rather than effectively remembering learned patterns. To address the challenges of forgetting and overfitting, we propose a novel method based on embedding space separation and compaction. Our method alleviates forgetting of previously learned tasks by forcing the feature distribution of new data away from the previous embedding space. It also mitigates overfitting by a memory calibration mechanism that encourages memory data to be close to its prototype to enhance intra-class compactness. In addition, the learnable parameters of the new task are initialized by drawing upon acquired knowledge from the previously learned task to facilitate forward knowledge transfer. With extensive experiments, we demonstrate that our method can significantly outperform previous state-of-the-art approaches.Comment: NAACL 2024 main conferenc

    Can ChatGPT-like Generative Models Guarantee Factual Accuracy? On the Mistakes of New Generation Search Engines

    Full text link
    Although large conversational AI models such as OpenAI's ChatGPT have demonstrated great potential, we question whether such models can guarantee factual accuracy. Recently, technology companies such as Microsoft and Google have announced new services which aim to combine search engines with conversational AI. However, we have found numerous mistakes in the public demonstrations that suggest we should not easily trust the factual claims of the AI models. Rather than criticizing specific models or companies, we hope to call on researchers and developers to improve AI models' transparency and factual correctness

    Language Models can be Logical Solvers

    Full text link
    Logical reasoning is a fundamental aspect of human intelligence and a key component of tasks like problem-solving and decision-making. Recent advancements have enabled Large Language Models (LLMs) to potentially exhibit reasoning capabilities, but complex logical reasoning remains a challenge. The state-of-the-art, solver-augmented language models, use LLMs to parse natural language logical questions into symbolic representations first and then adopt external logical solvers to take in the symbolic representations and output the answers. Despite their impressive performance, any parsing errors will inevitably result in the failure of the execution of the external logical solver and no answer to the logical questions. In this paper, we introduce LoGiPT, a novel language model that directly emulates the reasoning processes of logical solvers and bypasses the parsing errors by learning to strict adherence to solver syntax and grammar. LoGiPT is fine-tuned on a newly constructed instruction-tuning dataset derived from revealing and refining the invisible reasoning process of deductive solvers. Experimental results on two public deductive reasoning datasets demonstrate that LoGiPT outperforms state-of-the-art solver-augmented LMs and few-shot prompting methods on competitive LLMs like ChatGPT or GPT-4.Comment: Preprin

    Evaluation of a clinical pharmacist-led antimicrobial stewardship program in a neurosurgical intensive care unit: a pre-and post-intervention cohort study

    Get PDF
    Background: Antimicrobial resistance poses a significant challenge in neurosurgical intensive care units (ICU). The excessive use of broad-spectrum antibiotics is closely linked to the emergence and dissemination of drug-resistant bacteria within neurosurgical ICUs. This study assessed the effects of implementing a comprehensive Antimicrobial Stewardship (AMS) program in a neurosurgical ICU setting.Methods: From April 2022 to September 2022, an AMS program was implemented in the neurosurgical ICU. The program involved the regular presence of a pharmacist and an infectious disease physician who conducted prospective audits and provided feedback. To assess the impact of the AMS program, the outcome measures were compared between the AMS period and the 6 months before AMS implementation (pre-AMS period). The primary outcome was the use of antibacterial agents, including anti-pseudomonal beta-lactams (APBLs), polymyxin, and tigecycline. Additionally, the study evaluated the appropriateness of antimicrobial de-escalation and the susceptibility of Gram-negative bacilli to antimicrobial agents.Results: A total of 526 were included during the AMS period, while 487 patients were included in the pre-AMS period. The two groups had no significant differences in disease severity and mortality rates. During the AMS period, there was a notable decrease in the use of APBLs as empiric treatment (43.92% vs. 60.99%, p < 0.001). Multi-drug resistant organism (MDRO) infections decrease significantly during AMS period (11.03% vs. 18.48%, p < 0.001). The number of prescription adjustment increased significantly in all patients (0 item vs. 0 item, p < 0.001) and MDRO-positive patients (3 items vs. 2 items, p < 0.001) during the AMS period. Additionally, appropriate antimicrobial de-escalation for patients with MDRO showed improvement during the AMS period (39.66% vs. 20%, p = 0.001). Polymyxin utilization also decreased during the AMS period (15.52% vs. 31.11%, p = 0.034). Furthermore, the susceptibility of Gram-negative Bacilli isolates to APBLs was significantly higher during the AMS period.Conclusion: Implementing a comprehensive pharmacist-led AMS program led to a decrease in the use of antibacterial agents. This reduction in usage is significant because it can potentially delay the emergence of bacterial resistance

    Chain-of-Knowledge: Grounding Large Language Models via Dynamic Knowledge Adapting over Heterogeneous Sources

    Full text link
    We present chain-of-knowledge (CoK), a novel framework that augments large language models (LLMs) by dynamically incorporating grounding information from heterogeneous sources. It results in more factual rationales and reduced hallucination in generation. Specifically, CoK consists of three stages: reasoning preparation, dynamic knowledge adapting, and answer consolidation. Given a knowledge-intensive question, CoK first prepares several preliminary rationales and answers while identifying the relevant knowledge domains. If there is no majority consensus among the answers from samples, CoK corrects the rationales step by step by adapting knowledge from the identified domains. These corrected rationales can plausibly serve as a better foundation for the final answer consolidation. Unlike prior studies that primarily use unstructured data, CoK also leverages structured knowledge sources such as Wikidata and tables that provide more reliable factual information. To access both unstructured and structured knowledge sources in the dynamic knowledge adapting stage, we propose an adaptive query generator that allows the generation of queries for various types of query languages, including SPARQL, SQL, and natural sentences. Moreover, to minimize error propagation between rationales, CoK corrects the rationales progressively using preceding corrected rationales to generate and correct subsequent rationales. Extensive experiments show that CoK consistently improves the performance of LLMs on knowledge-intensive tasks across different domains.Comment: Accepted by ICLR 202

    ChatGPT's One-year Anniversary: Are Open-Source Large Language Models Catching up?

    Full text link
    Upon its release in late 2022, ChatGPT has brought a seismic shift in the entire landscape of AI, both in research and commerce. Through instruction-tuning a large language model (LLM) with supervised fine-tuning and reinforcement learning from human feedback, it showed that a model could answer human questions and follow instructions on a broad panel of tasks. Following this success, interests in LLMs have intensified, with new LLMs flourishing at frequent interval across academia and industry, including many start-ups focused on LLMs. While closed-source LLMs (e.g., OpenAI's GPT, Anthropic's Claude) generally outperform their open-source counterparts, the progress on the latter has been rapid with claims of achieving parity or even better on certain tasks. This has crucial implications not only on research but also on business. In this work, on the first anniversary of ChatGPT, we provide an exhaustive overview of this success, surveying all tasks where an open-source LLM has claimed to be on par or better than ChatGPT.Comment: version v4, included latest top-performing open-sourced LLM

    Retrieving Multimodal Information for Augmented Generation: A Survey

    Full text link
    As Large Language Models (LLMs) become popular, there emerged an important trend of using multimodality to augment the LLMs' generation ability, which enables LLMs to better interact with the world. However, there lacks a unified perception of at which stage and how to incorporate different modalities. In this survey, we review methods that assist and augment generative models by retrieving multimodal knowledge, whose formats range from images, codes, tables, graphs, to audio. Such methods offer a promising solution to important concerns such as factuality, reasoning, interpretability, and robustness. By providing an in-depth review, this survey is expected to provide scholars with a deeper understanding of the methods' applications and encourage them to adapt existing techniques to the fast-growing field of LLMs
    corecore