289 research outputs found
The role of electrochemical properties of biochar to promote methane production in anaerobic digestion
The electrochemical properties of biochar may be the key factor to promote anaerobic digestion, which has attracted extensive attention. However, the mechanism and the role of the electrochemical properties of biochar are remaining unclear. In this study, biochar with different electrochemical properties was prepared by pyrolysis at different temperatures (BC300/600/900) and oxidation or reduction modification (O/RBC300/600/900). The biochar was added as an additive to promote methanogenic performance of anaerobic digesters of glucose and food waste. In both anaerobic digestion systems, the cumulative methane production of food waste increased by 42.07% and the maximum methane production rate of glucose enhanced by 17.80% after BC900 treatment. RBC600 was inferior to BC900, but superior to BC600. Microbiological analysis suggests that biochar enriched the relative abundant Synergistia and Methanoculleus. This is conducive to the establishment of the direct interspecies electrons transfer (DIET). Results from correlation analysis, principal component analysis and machine learning confirmed that both of the electron donating capacities (EDC) and electrical conductivity (EC) are dominated factors affecting the cumulative methane yield. Through the analysis of electrochemical properties and preparation process of biochar, the results showed that the pyrolysis temperature increases and the content of phenolic hydroxyl decreases under medium temperature of biochar, which was beneficial to the methane production. This study found the key factors of the electrochemical properties of biochar in anaerobic digestion, provided new insights for the mechanism of biochar promoting anaerobic digestion and proposed novel directions for the preparation of biochar.acceptedVersio
DONAPI: Malicious NPM Packages Detector using Behavior Sequence Knowledge Mapping
With the growing popularity of modularity in software development comes the
rise of package managers and language ecosystems. Among them, npm stands out as
the most extensive package manager, hosting more than 2 million third-party
open-source packages that greatly simplify the process of building code.
However, this openness also brings security risks, as evidenced by numerous
package poisoning incidents.
In this paper, we synchronize a local package cache containing more than 3.4
million packages in near real-time to give us access to more package code
details. Further, we perform manual inspection and API call sequence analysis
on packages collected from public datasets and security reports to build a
hierarchical classification framework and behavioral knowledge base covering
different sensitive behaviors. In addition, we propose the DONAPI, an automatic
malicious npm packages detector that combines static and dynamic analysis. It
makes preliminary judgments on the degree of maliciousness of packages by code
reconstruction techniques and static analysis, extracts dynamic API call
sequences to confirm and identify obfuscated content that static analysis can
not handle alone, and finally tags malicious software packages based on the
constructed behavior knowledge base. To date, we have identified and manually
confirmed 325 malicious samples and discovered 2 unusual API calls and 246 API
call sequences that have not appeared in known samples.Comment: 18 pages, accepted for publication at USENIX Security 202
STU-Net: Scalable and Transferable Medical Image Segmentation Models Empowered by Large-Scale Supervised Pre-training
Large-scale models pre-trained on large-scale datasets have profoundly
advanced the development of deep learning. However, the state-of-the-art models
for medical image segmentation are still small-scale, with their parameters
only in the tens of millions. Further scaling them up to higher orders of
magnitude is rarely explored. An overarching goal of exploring large-scale
models is to train them on large-scale medical segmentation datasets for better
transfer capacities. In this work, we design a series of Scalable and
Transferable U-Net (STU-Net) models, with parameter sizes ranging from 14
million to 1.4 billion. Notably, the 1.4B STU-Net is the largest medical image
segmentation model to date. Our STU-Net is based on nnU-Net framework due to
its popularity and impressive performance. We first refine the default
convolutional blocks in nnU-Net to make them scalable. Then, we empirically
evaluate different scaling combinations of network depth and width, discovering
that it is optimal to scale model depth and width together. We train our
scalable STU-Net models on a large-scale TotalSegmentator dataset and find that
increasing model size brings a stronger performance gain. This observation
reveals that a large model is promising in medical image segmentation.
Furthermore, we evaluate the transferability of our model on 14 downstream
datasets for direct inference and 3 datasets for further fine-tuning, covering
various modalities and segmentation targets. We observe good performance of our
pre-trained model in both direct inference and fine-tuning. The code and
pre-trained models are available at https://github.com/Ziyan-Huang/STU-Net
A-Eval: A Benchmark for Cross-Dataset Evaluation of Abdominal Multi-Organ Segmentation
Although deep learning have revolutionized abdominal multi-organ
segmentation, models often struggle with generalization due to training on
small, specific datasets. With the recent emergence of large-scale datasets,
some important questions arise: \textbf{Can models trained on these datasets
generalize well on different ones? If yes/no, how to further improve their
generalizability?} To address these questions, we introduce A-Eval, a benchmark
for the cross-dataset Evaluation ('Eval') of Abdominal ('A') multi-organ
segmentation. We employ training sets from four large-scale public datasets:
FLARE22, AMOS, WORD, and TotalSegmentator, each providing extensive labels for
abdominal multi-organ segmentation. For evaluation, we incorporate the
validation sets from these datasets along with the training set from the BTCV
dataset, forming a robust benchmark comprising five distinct datasets. We
evaluate the generalizability of various models using the A-Eval benchmark,
with a focus on diverse data usage scenarios: training on individual datasets
independently, utilizing unlabeled data via pseudo-labeling, mixing different
modalities, and joint training across all available datasets. Additionally, we
explore the impact of model sizes on cross-dataset generalizability. Through
these analyses, we underline the importance of effective data usage in
enhancing models' generalization capabilities, offering valuable insights for
assembling large-scale datasets and improving training strategies. The code and
pre-trained models are available at
\href{https://github.com/uni-medical/A-Eval}{https://github.com/uni-medical/A-Eval}
SA-Med2D-20M Dataset: Segment Anything in 2D Medical Imaging with 20 Million masks
Segment Anything Model (SAM) has achieved impressive results for natural
image segmentation with input prompts such as points and bounding boxes. Its
success largely owes to massive labeled training data. However, directly
applying SAM to medical image segmentation cannot perform well because SAM
lacks medical knowledge -- it does not use medical images for training. To
incorporate medical knowledge into SAM, we introduce SA-Med2D-20M, a
large-scale segmentation dataset of 2D medical images built upon numerous
public and private datasets. It consists of 4.6 million 2D medical images and
19.7 million corresponding masks, covering almost the whole body and showing
significant diversity. This paper describes all the datasets collected in
SA-Med2D-20M and details how to process these datasets. Furthermore,
comprehensive statistics of SA-Med2D-20M are presented to facilitate the better
use of our dataset, which can help the researchers build medical vision
foundation models or apply their models to downstream medical applications. We
hope that the large scale and diversity of SA-Med2D-20M can be leveraged to
develop medical artificial intelligence for enhancing diagnosis, medical image
analysis, knowledge sharing, and education. The data with the redistribution
license is publicly available at https://github.com/OpenGVLab/SAM-Med2D
Magnetic resonance imaging based deep-learning model: a rapid, high-performance, automated tool for testicular volume measurements
BackgroundTesticular volume (TV) is an essential parameter for monitoring testicular functions and pathologies. Nevertheless, current measurement tools, including orchidometers and ultrasonography, encounter challenges in obtaining accurate and personalized TV measurements.PurposeBased on magnetic resonance imaging (MRI), this study aimed to establish a deep learning model and evaluate its efficacy in segmenting the testes and measuring TV.Materials and methodsThe study cohort consisted of retrospectively collected patient data (N = 200) and a prospectively collected dataset comprising 10 healthy volunteers. The retrospective dataset was divided into training and independent validation sets, with an 8:2 random distribution. Each of the 10 healthy volunteers underwent 5 scans (forming the testing dataset) to evaluate the measurement reproducibility. A ResUNet algorithm was applied to segment the testes. Volume of each testis was calculated by multiplying the voxel volume by the number of voxels. Manually determined masks by experts were used as ground truth to assess the performance of the deep learning model.ResultsThe deep learning model achieved a mean Dice score of 0.926 ± 0.034 (0.921 ± 0.026 for the left testis and 0.926 ± 0.034 for the right testis) in the validation cohort and a mean Dice score of 0.922 ± 0.02 (0.931 ± 0.019 for the left testis and 0.932 ± 0.022 for the right testis) in the testing cohort. There was strong correlation between the manual and automated TV (R2 ranging from 0.974 to 0.987 in the validation cohort; R2 ranging from 0.936 to 0.973 in the testing cohort). The volume differences between the manual and automated measurements were 0.838 ± 0.991 (0.209 ± 0.665 for LTV and 0.630 ± 0.728 for RTV) in the validation cohort and 0.815 ± 0.824 (0.303 ± 0.664 for LTV and 0.511 ± 0.444 for RTV) in the testing cohort. Additionally, the deep-learning model exhibited excellent reproducibility (intraclass correlation >0.9) in determining TV.ConclusionThe MRI-based deep learning model is an accurate and reliable tool for measuring TV
Potential Diagnostic Applications of Multi-Delay Arterial Spin Labeling in Early Alzheimer’s Disease: The Chinese Imaging, Biomarkers, and Lifestyle Study
Background: Cerebral blood flow (CBF) alterations are involved in the onset and progression of Alzheimer’s disease (AD) and can be a potential biomarker. However, CBF measured by single-delay arterial spin labeling (ASL) for discrimination of mild cognitive impairment (MCI, an early stage of AD) was lack of accuracy. Multi-delay ASL can not only provide CBF quantification but also provide arterial transit time (ATT). Unfortunately, the technique was scarcely applied to the diagnosis of AD. Here, we detected the utility of ASL with 1-delay and 7-delay in ten regions of interest (ROIs) to identify MCI and AD. Materials and Methods: Pseudocontinuous ASL (pCASL) MRI was acquired on a 3T GE scanner in adults from the Chinese Imaging, Biomarkers, and Lifestyle (CIBL) Study of AD cohort, including 26 normal cognition (NC), 37 MCI, and 39 AD. Receiver operating characteristic (ROC) analyses with 1-delay and 7-delay ASL were performed for the identification of MCI and AD. The DeLong test was used to compare ROC curves. Results: For CBF of 1-delay or 7-delay the AUCs showed moderate-high performance for the AD/NC and AD/MCI comparisons (AUC = 0.83∼0.96) (p 0.05). Conclusion: The combination of CBF and ATT with 7-delay ASL showed higher performance for identification of MCI than CBF of 1-delay, when adding to sex, age, APOE ε4 carrier status, and education years, the diagnostic performance was further increased, presenting a potential imaging biomarker in early AD
- …