289 research outputs found

    The role of electrochemical properties of biochar to promote methane production in anaerobic digestion

    Get PDF
    The electrochemical properties of biochar may be the key factor to promote anaerobic digestion, which has attracted extensive attention. However, the mechanism and the role of the electrochemical properties of biochar are remaining unclear. In this study, biochar with different electrochemical properties was prepared by pyrolysis at different temperatures (BC300/600/900) and oxidation or reduction modification (O/RBC300/600/900). The biochar was added as an additive to promote methanogenic performance of anaerobic digesters of glucose and food waste. In both anaerobic digestion systems, the cumulative methane production of food waste increased by 42.07% and the maximum methane production rate of glucose enhanced by 17.80% after BC900 treatment. RBC600 was inferior to BC900, but superior to BC600. Microbiological analysis suggests that biochar enriched the relative abundant Synergistia and Methanoculleus. This is conducive to the establishment of the direct interspecies electrons transfer (DIET). Results from correlation analysis, principal component analysis and machine learning confirmed that both of the electron donating capacities (EDC) and electrical conductivity (EC) are dominated factors affecting the cumulative methane yield. Through the analysis of electrochemical properties and preparation process of biochar, the results showed that the pyrolysis temperature increases and the content of phenolic hydroxyl decreases under medium temperature of biochar, which was beneficial to the methane production. This study found the key factors of the electrochemical properties of biochar in anaerobic digestion, provided new insights for the mechanism of biochar promoting anaerobic digestion and proposed novel directions for the preparation of biochar.acceptedVersio

    DONAPI: Malicious NPM Packages Detector using Behavior Sequence Knowledge Mapping

    Full text link
    With the growing popularity of modularity in software development comes the rise of package managers and language ecosystems. Among them, npm stands out as the most extensive package manager, hosting more than 2 million third-party open-source packages that greatly simplify the process of building code. However, this openness also brings security risks, as evidenced by numerous package poisoning incidents. In this paper, we synchronize a local package cache containing more than 3.4 million packages in near real-time to give us access to more package code details. Further, we perform manual inspection and API call sequence analysis on packages collected from public datasets and security reports to build a hierarchical classification framework and behavioral knowledge base covering different sensitive behaviors. In addition, we propose the DONAPI, an automatic malicious npm packages detector that combines static and dynamic analysis. It makes preliminary judgments on the degree of maliciousness of packages by code reconstruction techniques and static analysis, extracts dynamic API call sequences to confirm and identify obfuscated content that static analysis can not handle alone, and finally tags malicious software packages based on the constructed behavior knowledge base. To date, we have identified and manually confirmed 325 malicious samples and discovered 2 unusual API calls and 246 API call sequences that have not appeared in known samples.Comment: 18 pages, accepted for publication at USENIX Security 202

    STU-Net: Scalable and Transferable Medical Image Segmentation Models Empowered by Large-Scale Supervised Pre-training

    Full text link
    Large-scale models pre-trained on large-scale datasets have profoundly advanced the development of deep learning. However, the state-of-the-art models for medical image segmentation are still small-scale, with their parameters only in the tens of millions. Further scaling them up to higher orders of magnitude is rarely explored. An overarching goal of exploring large-scale models is to train them on large-scale medical segmentation datasets for better transfer capacities. In this work, we design a series of Scalable and Transferable U-Net (STU-Net) models, with parameter sizes ranging from 14 million to 1.4 billion. Notably, the 1.4B STU-Net is the largest medical image segmentation model to date. Our STU-Net is based on nnU-Net framework due to its popularity and impressive performance. We first refine the default convolutional blocks in nnU-Net to make them scalable. Then, we empirically evaluate different scaling combinations of network depth and width, discovering that it is optimal to scale model depth and width together. We train our scalable STU-Net models on a large-scale TotalSegmentator dataset and find that increasing model size brings a stronger performance gain. This observation reveals that a large model is promising in medical image segmentation. Furthermore, we evaluate the transferability of our model on 14 downstream datasets for direct inference and 3 datasets for further fine-tuning, covering various modalities and segmentation targets. We observe good performance of our pre-trained model in both direct inference and fine-tuning. The code and pre-trained models are available at https://github.com/Ziyan-Huang/STU-Net

    A-Eval: A Benchmark for Cross-Dataset Evaluation of Abdominal Multi-Organ Segmentation

    Full text link
    Although deep learning have revolutionized abdominal multi-organ segmentation, models often struggle with generalization due to training on small, specific datasets. With the recent emergence of large-scale datasets, some important questions arise: \textbf{Can models trained on these datasets generalize well on different ones? If yes/no, how to further improve their generalizability?} To address these questions, we introduce A-Eval, a benchmark for the cross-dataset Evaluation ('Eval') of Abdominal ('A') multi-organ segmentation. We employ training sets from four large-scale public datasets: FLARE22, AMOS, WORD, and TotalSegmentator, each providing extensive labels for abdominal multi-organ segmentation. For evaluation, we incorporate the validation sets from these datasets along with the training set from the BTCV dataset, forming a robust benchmark comprising five distinct datasets. We evaluate the generalizability of various models using the A-Eval benchmark, with a focus on diverse data usage scenarios: training on individual datasets independently, utilizing unlabeled data via pseudo-labeling, mixing different modalities, and joint training across all available datasets. Additionally, we explore the impact of model sizes on cross-dataset generalizability. Through these analyses, we underline the importance of effective data usage in enhancing models' generalization capabilities, offering valuable insights for assembling large-scale datasets and improving training strategies. The code and pre-trained models are available at \href{https://github.com/uni-medical/A-Eval}{https://github.com/uni-medical/A-Eval}

    SA-Med2D-20M Dataset: Segment Anything in 2D Medical Imaging with 20 Million masks

    Full text link
    Segment Anything Model (SAM) has achieved impressive results for natural image segmentation with input prompts such as points and bounding boxes. Its success largely owes to massive labeled training data. However, directly applying SAM to medical image segmentation cannot perform well because SAM lacks medical knowledge -- it does not use medical images for training. To incorporate medical knowledge into SAM, we introduce SA-Med2D-20M, a large-scale segmentation dataset of 2D medical images built upon numerous public and private datasets. It consists of 4.6 million 2D medical images and 19.7 million corresponding masks, covering almost the whole body and showing significant diversity. This paper describes all the datasets collected in SA-Med2D-20M and details how to process these datasets. Furthermore, comprehensive statistics of SA-Med2D-20M are presented to facilitate the better use of our dataset, which can help the researchers build medical vision foundation models or apply their models to downstream medical applications. We hope that the large scale and diversity of SA-Med2D-20M can be leveraged to develop medical artificial intelligence for enhancing diagnosis, medical image analysis, knowledge sharing, and education. The data with the redistribution license is publicly available at https://github.com/OpenGVLab/SAM-Med2D

    Magnetic resonance imaging based deep-learning model: a rapid, high-performance, automated tool for testicular volume measurements

    Get PDF
    BackgroundTesticular volume (TV) is an essential parameter for monitoring testicular functions and pathologies. Nevertheless, current measurement tools, including orchidometers and ultrasonography, encounter challenges in obtaining accurate and personalized TV measurements.PurposeBased on magnetic resonance imaging (MRI), this study aimed to establish a deep learning model and evaluate its efficacy in segmenting the testes and measuring TV.Materials and methodsThe study cohort consisted of retrospectively collected patient data (N = 200) and a prospectively collected dataset comprising 10 healthy volunteers. The retrospective dataset was divided into training and independent validation sets, with an 8:2 random distribution. Each of the 10 healthy volunteers underwent 5 scans (forming the testing dataset) to evaluate the measurement reproducibility. A ResUNet algorithm was applied to segment the testes. Volume of each testis was calculated by multiplying the voxel volume by the number of voxels. Manually determined masks by experts were used as ground truth to assess the performance of the deep learning model.ResultsThe deep learning model achieved a mean Dice score of 0.926 ± 0.034 (0.921 ± 0.026 for the left testis and 0.926 ± 0.034 for the right testis) in the validation cohort and a mean Dice score of 0.922 ± 0.02 (0.931 ± 0.019 for the left testis and 0.932 ± 0.022 for the right testis) in the testing cohort. There was strong correlation between the manual and automated TV (R2 ranging from 0.974 to 0.987 in the validation cohort; R2 ranging from 0.936 to 0.973 in the testing cohort). The volume differences between the manual and automated measurements were 0.838 ± 0.991 (0.209 ± 0.665 for LTV and 0.630 ± 0.728 for RTV) in the validation cohort and 0.815 ± 0.824 (0.303 ± 0.664 for LTV and 0.511 ± 0.444 for RTV) in the testing cohort. Additionally, the deep-learning model exhibited excellent reproducibility (intraclass correlation >0.9) in determining TV.ConclusionThe MRI-based deep learning model is an accurate and reliable tool for measuring TV

    Potential Diagnostic Applications of Multi-Delay Arterial Spin Labeling in Early Alzheimer’s Disease: The Chinese Imaging, Biomarkers, and Lifestyle Study

    Get PDF
    Background: Cerebral blood flow (CBF) alterations are involved in the onset and progression of Alzheimer’s disease (AD) and can be a potential biomarker. However, CBF measured by single-delay arterial spin labeling (ASL) for discrimination of mild cognitive impairment (MCI, an early stage of AD) was lack of accuracy. Multi-delay ASL can not only provide CBF quantification but also provide arterial transit time (ATT). Unfortunately, the technique was scarcely applied to the diagnosis of AD. Here, we detected the utility of ASL with 1-delay and 7-delay in ten regions of interest (ROIs) to identify MCI and AD. Materials and Methods: Pseudocontinuous ASL (pCASL) MRI was acquired on a 3T GE scanner in adults from the Chinese Imaging, Biomarkers, and Lifestyle (CIBL) Study of AD cohort, including 26 normal cognition (NC), 37 MCI, and 39 AD. Receiver operating characteristic (ROC) analyses with 1-delay and 7-delay ASL were performed for the identification of MCI and AD. The DeLong test was used to compare ROC curves. Results: For CBF of 1-delay or 7-delay the AUCs showed moderate-high performance for the AD/NC and AD/MCI comparisons (AUC = 0.83∼0.96) (p 0.05). Conclusion: The combination of CBF and ATT with 7-delay ASL showed higher performance for identification of MCI than CBF of 1-delay, when adding to sex, age, APOE ε4 carrier status, and education years, the diagnostic performance was further increased, presenting a potential imaging biomarker in early AD
    • …
    corecore