25 research outputs found

    Question Decomposition Tree for Answering Complex Questions over Knowledge Bases

    Full text link
    Knowledge base question answering (KBQA) has attracted a lot of interest in recent years, especially for complex questions which require multiple facts to answer. Question decomposition is a promising way to answer complex questions. Existing decomposition methods split the question into sub-questions according to a single compositionality type, which is not sufficient for questions involving multiple compositionality types. In this paper, we propose Question Decomposition Tree (QDT) to represent the structure of complex questions. Inspired by recent advances in natural language generation (NLG), we present a two-staged method called Clue-Decipher to generate QDT. It can leverage the strong ability of NLG model and simultaneously preserve the original questions. To verify that QDT can enhance KBQA task, we design a decomposition-based KBQA system called QDTQA. Extensive experiments show that QDTQA outperforms previous state-of-the-art methods on ComplexWebQuestions dataset. Besides, our decomposition method improves an existing KBQA system by 12% and sets a new state-of-the-art on LC-QuAD 1.0.Comment: Accepted by AAAI202

    TRANSOM: An Efficient Fault-Tolerant System for Training LLMs

    Full text link
    Large language models (LLMs) with hundreds of billions or trillions of parameters, represented by chatGPT, have achieved profound impact on various fields. However, training LLMs with super-large-scale parameters requires large high-performance GPU clusters and long training periods lasting for months. Due to the inevitable hardware and software failures in large-scale clusters, maintaining uninterrupted and long-duration training is extremely challenging. As a result, A substantial amount of training time is devoted to task checkpoint saving and loading, task rescheduling and restart, and task manual anomaly checks, which greatly harms the overall training efficiency. To address these issues, we propose TRANSOM, a novel fault-tolerant LLM training system. In this work, we design three key subsystems: the training pipeline automatic fault tolerance and recovery mechanism named Transom Operator and Launcher (TOL), the training task multi-dimensional metric automatic anomaly detection system named Transom Eagle Eye (TEE), and the training checkpoint asynchronous access automatic fault tolerance and recovery technology named Transom Checkpoint Engine (TCE). Here, TOL manages the lifecycle of training tasks, while TEE is responsible for task monitoring and anomaly reporting. TEE detects training anomalies and reports them to TOL, who automatically enters the fault tolerance strategy to eliminate abnormal nodes and restart the training task. And the asynchronous checkpoint saving and loading functionality provided by TCE greatly shorten the fault tolerance overhead. The experimental results indicate that TRANSOM significantly enhances the efficiency of large-scale LLM training on clusters. Specifically, the pre-training time for GPT3-175B has been reduced by 28%, while checkpoint saving and loading performance have improved by a factor of 20.Comment: 14 pages, 9 figure

    MRI radiomics-based decision support tool for a personalized classification of cervical disc degeneration: a two-center study

    Get PDF
    Objectives: To develop and validate an MRI radiomics-based decision support tool for the automated grading of cervical disc degeneration.Methods: The retrospective study included 2,610 cervical disc samples of 435 patients from two hospitals. The cervical magnetic resonance imaging (MRI) analysis of patients confirmed cervical disc degeneration grades using the Pfirrmann grading system. A training set (1,830 samples of 305 patients) and an independent test set (780 samples of 130 patients) were divided for the construction and validation of the machine learning model, respectively. We provided a fine-tuned MedSAM model for automated cervical disc segmentation. Then, we extracted 924 radiomic features from each segmented disc in T1 and T2 MRI modalities. All features were processed and selected using minimum redundancy maximum relevance (mRMR) and multiple machine learning algorithms. Meanwhile, the radiomics models of various machine learning algorithms and MRI images were constructed and compared. Finally, the combined radiomics model was constructed in the training set and validated in the test set. Radiomic feature mapping was provided for auxiliary diagnosis.Results: Of the 2,610 cervical disc samples, 794 (30.4%) were classified as low grade and 1,816 (69.6%) were classified as high grade. The fine-tuned MedSAM model achieved good segmentation performance, with the mean Dice coefficient of 0.93. Higher-order texture features contributed to the dominant force in the diagnostic task (80%). Among various machine learning models, random forest performed better than the other algorithms (p < 0.01), and the T2 MRI radiomics model showed better results than T1 MRI in the diagnostic performance (p < 0.05). The final combined radiomics model had an area under the receiver operating characteristic curve (AUC) of 0.95, an accuracy of 89.51%, a precision of 87.07%, a recall of 98.83%, and an F1 score of 0.93 in the test set, which were all better than those of other models (p < 0.05).Conclusion: The radiomics-based decision support tool using T1 and T2 MRI modalities can be used for cervical disc degeneration grading, facilitating individualized management

    Question Decomposition Tree for Answering Complex Questions over Knowledge Bases

    No full text
    Knowledge base question answering (KBQA) has attracted a lot of interest in recent years, especially for complex questions which require multiple facts to answer. Question decomposition is a promising way to answer complex questions. Existing decomposition methods split the question into sub-questions according to a single compositionality type, which is not sufficient for questions involving multiple compositionality types. In this paper, we propose Question Decomposition Tree (QDT) to represent the structure of complex questions. Inspired by recent advances in natural language generation (NLG), we present a two-staged method called Clue-Decipher to generate QDT. It can leverage the strong ability of NLG model and simultaneously preserve the original questions. To verify that QDT can enhance KBQA task, we design a decomposition-based KBQA system called QDTQA. Extensive experiments show that QDTQA outperforms previous state-of-the-art methods on ComplexWebQuestions dataset. Besides, our decomposition method improves an existing KBQA system by 12% and sets a new state-of-the-art on LC-QuAD 1.0

    Acoustic Indoor Localization System Integrating TDMA+FDMA Transmission Scheme and Positioning Correction Technique

    No full text
    This paper presents a novel audio indoor localization system. In the proposed system, four speakers placed at known positions transmit chirp signals according to the time-division multiple access (TDMA) plus frequency-division multiple access (FDMA) transmission scheme. A smartphone receives the signal via a built-in microphone and calculates the time differences of arrival (TDOAs). Using TDOA measurements, the position is estimated by the shrinking-circle method. In particular, to reduce the positioning error in moving conditions, a TDOA correction method based on Doppler shifts is proposed. The performance of the proposed system was evaluated in real-world experiments using a 10.971 m × 5.684 m positioning area. The results of the static-target positioning experiment showed that the TDMA+FDMA transmission scheme has more advantages in improving the update rate of the positioning system than the TDMA-only transmission scheme. The results of the moving-target positioning experiment under three different speeds demonstrated that the positioning errors were reduced by about 10 cm when the Doppler-shift-based TDOA correction method was adopted. This research provides a possible framework for the realization of a TDOA-chirp-based acoustic indoor positioning system with high positioning accuracy and update rate

    Effect of detoxification methods on ABE production from corn stover hydrolysate by Clostridium acetobutylicum CICC 8016

    No full text
    In this study, effects of different single biomass derived inhibitors on acetone–butanol–ethanol (ABE) production by Clostridium acetobutylicum CICC 8016 were first investigated. The results showed that formic acid, coumaric acid, and furfural at 0.5 g/L (sodium formate equivalent) inhibited ABE production. Furthermore, corn stover hydrolysate media were prepared following dilute acid pretreatment, enzymatic hydrolysis, and detoxification with different methods. Among overliming, steam stripping, acetone–ethyl ether extraction, and ion exchange with five anion resins, adsorption with resin D301 showed the highest efficiency for inhibitor removal (99–100% of phenolics and 87–99% of sugar degradation products). Without detoxification, ABE production was lower than 1.0 g/L from 28.1 g/L sugars whereas ABE production with medium detoxified by D301 resin achieved higher ABE concentrations and yields than control with synthetic medium. Correlation analysis further revealed that formic acid, coumaric acid, and total phenolics were the major compounds inhibiting ABE production. The results also showed that the single detoxification method was sufficient to detoxify the hydrolysate for ABE production at the pretreatment conditions used in this study.</p

    A study of the kinematic characteristics and energy conversion of waves generated by granular landslide

    No full text
    The consequences of landslide waves are far beyond the landslide itself that has attracted widespread attention. The prediction and evaluation of this kind of disaster has always been difficult, and the precise description of landslide surge motion characteristics and energy conversion law is the key and premise to solve the problem. In this paper, we use gravel to mimic granular landslide and establish a 3D landslide wave model in a rectangular flume, aiming to analyze how the landslide volume, velocity and water depth affect landslide accumulation, wave characteristics and energy conversion. The results show that (1) the waves generated by solid impacting the water are affected by the landslide size and shape. Slides with larger thickness and faster speed tend to produce nonlinear transition wave, and thin and slow slides generally produce nonlinear oscillation wave. (2) The volume effect and velocity effect based on the test reveal that the surge scale of a certain water depth is positively correlated with the landslide volume and velocity. The water depth effect explains the differences of waveform and velocity under a certain wave energy. Statistic results show that under the shallow water conditions, surge height on average is 67% higher and wave speed on average is 51.17% higher than those under the deep water conditions. (3) The conversion rate between landslide energy and wave energy ranges from 1.00% to 3.07%. 3D experiments encounter more energy dissipation due to diffusion and its conversion rate is lower than that in the 2D experiments. This study discusses the kinematic characteristics of granular landslide, first wave generation, propagation and inundation, and proposes the basic law of energy conversion between landslide and water. It id of certain value and significance for landslide wave hazard prevention and mitigation

    Knowledge Extraction and Quality Inspection of Chinese Petrographic Description Texts with Complex Entities and Relations Using Machine Reading and Knowledge Graph: A Preliminary Research Study

    No full text
    (1) Background: Geological surveying is undergoing a digital transformation process towards the adoption of intelligent methods in China. Cognitive intelligence methods, such as those based on knowledge graphs and machine reading, have made progress in many domains and also provide a technical basis for quality detection in unstructured lithographic description texts. (2) Methods: First, the named entities and the relations of the domain-specific knowledge graph of petrography were defined based on the petrographic theory. Second, research was carried out based on a manually annotated corpus of petrographic description. The extraction of N-ary and single-entity overlapping relations and the separation of complex entities are key steps in this process. Third, a petrographic knowledge graph was formulated based on prior knowledge. Finally, the consistency between knowledge triples extracted from the corpus and the petrographic knowledge graph was calculated. The 1:50,000 sheet of Fengxiangyi located in the Dabie orogenic belt was selected for the empirical research. (3) Results: Using machine reading and the knowledge graph, petrographic knowledge can be extracted and the knowledge consistency calculation can quickly detect description errors about textures, structures and mineral components in petrographic description. (4) Conclusions: The proposed framework can be used to realise the intelligent inspection of petrographic knowledge with complex entities and relations and to improve the quality of petrographic description texts effectively
    corecore