484,332 research outputs found
Deep learning for in vitro prediction of pharmaceutical formulations
Current pharmaceutical formulation development still strongly relies on the
traditional trial-and-error approach by individual experiences of
pharmaceutical scientists, which is laborious, time-consuming and costly.
Recently, deep learning has been widely applied in many challenging domains
because of its important capability of automatic feature extraction. The aim of
this research is to use deep learning to predict pharmaceutical formulations.
In this paper, two different types of dosage forms were chosen as model
systems. Evaluation criteria suitable for pharmaceutics were applied to
assessing the performance of the models. Moreover, an automatic dataset
selection algorithm was developed for selecting the representative data as
validation and test datasets. Six machine learning methods were compared with
deep learning. The result shows the accuracies of both two deep neural networks
were above 80% and higher than other machine learning models, which showed good
prediction in pharmaceutical formulations. In summary, deep learning with the
automatic data splitting algorithm and the evaluation criteria suitable for
pharmaceutical formulation data was firstly developed for the prediction of
pharmaceutical formulations. The cross-disciplinary integration of
pharmaceutics and artificial intelligence may shift the paradigm of
pharmaceutical researches from experience-dependent studies to data-driven
methodologies
Multitask Deep Learning for Accurate Risk Stratification and Prediction of Next Steps for Coronary CT Angiography Patients
Diagnostic investigation has an important role in risk stratification and
clinical decision making of patients with suspected and documented Coronary
Artery Disease (CAD). However, the majority of existing tools are primarily
focused on the selection of gatekeeper tests, whereas only a handful of systems
contain information regarding the downstream testing or treatment. We propose a
multi-task deep learning model to support risk stratification and down-stream
test selection for patients undergoing Coronary Computed Tomography Angiography
(CCTA). The analysis included 14,021 patients who underwent CCTA between 2006
and 2017. Our novel multitask deep learning framework extends the state-of-the
art Perceiver model to deal with real-world CCTA report data. Our model
achieved an Area Under the receiver operating characteristic Curve (AUC) of
0.76 in CAD risk stratification, and 0.72 AUC in predicting downstream tests.
Our proposed deep learning model can accurately estimate the likelihood of CAD
and provide recommended downstream tests based on prior CCTA data. In clinical
practice, the utilization of such an approach could bring a paradigm shift in
risk stratification and downstream management. Despite significant progress
using deep learning models for tabular data, they do not outperform gradient
boosting decision trees, and further research is required in this area.
However, neural networks appear to benefit more readily from multi-task
learning than tree-based models. This could offset the shortcomings of using
single task learning approach when working with tabular data
Evaluating the Robustness of Test Selection Methods for Deep Neural Networks
Testing deep learning-based systems is crucial but challenging due to the
required time and labor for labeling collected raw data. To alleviate the
labeling effort, multiple test selection methods have been proposed where only
a subset of test data needs to be labeled while satisfying testing
requirements. However, we observe that such methods with reported promising
results are only evaluated under simple scenarios, e.g., testing on original
test data. This brings a question to us: are they always reliable? In this
paper, we explore when and to what extent test selection methods fail for
testing. Specifically, first, we identify potential pitfalls of 11 selection
methods from top-tier venues based on their construction. Second, we conduct a
study on five datasets with two model architectures per dataset to empirically
confirm the existence of these pitfalls. Furthermore, we demonstrate how
pitfalls can break the reliability of these methods. Concretely, methods for
fault detection suffer from test data that are: 1) correctly classified but
uncertain, or 2) misclassified but confident. Remarkably, the test relative
coverage achieved by such methods drops by up to 86.85%. On the other hand,
methods for performance estimation are sensitive to the choice of
intermediate-layer output. The effectiveness of such methods can be even worse
than random selection when using an inappropriate layer.Comment: 12 page
On the integration of conceptual hierarchies with deep learning for explainable open-domain question answering
Question Answering, with its potential to make human-computer interactions more intuitive, has had a revival in recent years with the influx of deep learning methods into natural language processing and the simultaneous adoption of personal assistants such as Siri, Google Now, and Alexa. Unfortunately, Question Classification, an essential element of question answering, which classifies questions based on the class of the expected answer had been overlooked. Although the task of question classification was explicitly developed for use in question answering systems, the more advanced task of question classification, which classifies questions into between fifty and a hundred question classes, had developed into independent tasks with no application in question answering.
The work presented in this thesis bridges this gap by making use of fine-grained question classification for answer selection, arguably the most challenging subtask of question answering, and hence the defacto standard of measure of its performance on question answering. The use of question classification in a downstream task required significant improvement to question classification, which was achieved in this work by integrating linguistic information and deep learning through what we call Types, a novel method of representing Concepts.
Our work on a purely rule-based system for fine-grained Question Classification using Types achieved an accuracy of 97.2%, close to a 6 point improvement over the previous state of the art and has remained state of the art in question classification for over two years. The integration of these question classes and a deep learning model for Answer Selection resulted in MRR and MAP scores which outperform the current state of the art by between 3 and 5 points on both versions of a standard test set
Deep learning through the case method
[EN] Environmental Impact Assessment is a subject that aims to sensitize students about the need to study and adequately foresee the consequences that human actions have on the environment. Thus, this subject allows to address the syllabus in an interdisciplinary, complex and dynamic work environment. To success, it is essential that students achieve deep learning, and be able to identify patterns and connections in systems. For that, the subject must be developed as a whole. For this reason, the Case Method is chosen as the learning methodology, by facilitating the relationship with the reality of the selected cases and achieving in the students a greater capacity for analysis, interpretation and use of the concepts worked, enhancing their meaningful learning. Thus, the objective of this research was to verify whether the use of the case method, a methodology focused on learning, improved the learning strategies of students at the University level. For this, we used a pre-test/post-test experimental design using the CEVEAPEU questionnaire.
The results showed that students use more and better learning strategies. There are significant differences in the students' learning strategies, in the global score, in the two scales and four out of six subscales: Motivational strategies, Metacognitive strategies, Information search and selection strategies, and Processing and use strategies. The use of the case method as a pedagogical tool allowed students to learn better, both individually and in groups. This methodology required a proactive, constant and cooperative participation of the students, that promote the responsibility in their work development and allows to get closer to their professional future.Romero Gil, I.; Paches Giner, MAV. (2021). Deep learning through the case method. IATED Academy. 5527-5536. https://doi.org/10.21125/inted.2021.1118S5527553
Recommended from our members
Simulating drug responses in laboratory test time series with deep generative modeling
Drug effects can be unpredictable and vary widely among patients with environmental, genetic, and clinical factors. Randomized control trials (RCTs) are not sufficient to identify adverse drug reactions (ADRs), and the electronic health record (EHR) along with medical claims have become an important resource for pharmacovigilance. Among all the data collected in hospitals, laboratory tests represent the most documented and reliable data type in the EHR. Laboratory tests are at the core of the clinical decision process and are used for diagnosis, monitoring, screening, and research by physicians. They can be linked to drug effects either directly, with therapeutic drug monitoring (TDM), or indirectly using drug laboratory effects (DLEs) that affect surrogate tests. Unfortunately, very few automated methods use laboratory tests to inform clinical decision making and predict drug effects, partly due to the complexity of these time series that are irregularly sampled, highly dependent on other clinical covariates, and non-stationary.
Deep learning, the branch of machine learning that relies on high-capacity artificial neural networks, has known a renewed popularity this past decade and has transformed fields such as computer vision and natural language processing. Deep learning holds the promise of better performances compared to established machine learning models, although with the necessity for larger training datasets due to their higher degrees of freedom. These models are more flexible with multi-modal inputs and can make sense of large amounts of features without extensive engineering. Both qualities make deep learning models ideal candidate for complex, multi-modal, noisy healthcare datasets.
With the development of novel deep learning methods such as generative adversarial networks (GANs), there is an unprecedented opportunity to learn how to augment existing clinical dataset with realistic synthetic data and increase predictive performances. Moreover, GANs have the potential to simulate effects of individual covariates such as drug exposures by leveraging the properties of implicit generative models.
In this dissertation, I present a body of work that aims at paving the way for next generation laboratory test-based clinical decision support systems powered by deep learning. To this end, I organized my experiments around three building blocks: (1) the evaluation of various deep learning architectures with laboratory test time series and their covariates with a forecasting task; (2) the development of implicit generative models of laboratory test time series using the Wasserstein GAN framework; (3) the inference properties of these models for the simulation of drug effects in laboratory test time series, and their application for data augmentation. Each component has its own evaluation: The forecasting task enabled me to explore the properties and performances of different learning architectures; the Wasserstein GAN models are evaluated with both intrinsic metrics and extrinsic tasks, and I always set baselines to avoid providing results in a "neural-network only" referential. Applied machine learning, and more so with deep learning, is an empirical science. While the datasets used in this dissertation are not publicly available due to patient privacy regulation, I described pre-processing steps, hyper-parameters selection and training processes with reproducibility and transparency in mind.
In the specific context of these studies involving laboratory test time series and their clinical covariates, I found that for supervised tasks, machine learning holds up well against deep learning methods. Complex recurrent architectures like long short-term memory (LSTM) do not perform well on these short time series, while convolutional neural networks (CNNs) and multi-layer perceptrons (MLPs) provide the best performances, at the cost of extensive hyper-parameter tuning. Generative adversarial networks, enabled by deep learning models, were able to generate high-fidelity laboratory test time series, and the quality of the generated samples was increased with conditional models using drug exposures as auxiliary information. Interestingly, forecasting models trained on synthetic data exclusively still retain good performances, confirming the potential of GANs in privacy-oriented applications.
Finally, conditional GANs demonstrated an ability to interpolate samples from drug exposure combinations not seen during training, opening the way for laboratory test simulation with larger auxiliary information spaces. In specific cases, augmenting real training sets with synthetic data improved performances in the forecasting tasks, and could be extended to other applications where rare cases present a high prediction error
Meta-Learning Strategies through Value Maximization in Neural Networks
Biological and artificial learning agents face numerous choices about how to
learn, ranging from hyperparameter selection to aspects of task distributions
like curricula. Understanding how to make these meta-learning choices could
offer normative accounts of cognitive control functions in biological learners
and improve engineered systems. Yet optimal strategies remain challenging to
compute in modern deep networks due to the complexity of optimizing through the
entire learning process. Here we theoretically investigate optimal strategies
in a tractable setting. We present a learning effort framework capable of
efficiently optimizing control signals on a fully normative objective:
discounted cumulative performance throughout learning. We obtain computational
tractability by using average dynamical equations for gradient descent,
available for simple neural network architectures. Our framework accommodates a
range of meta-learning and automatic curriculum learning methods in a unified
normative setting. We apply this framework to investigate the effect of
approximations in common meta-learning algorithms; infer aspects of optimal
curricula; and compute optimal neuronal resource allocation in a continual
learning setting. Across settings, we find that control effort is most
beneficial when applied to easier aspects of a task early in learning; followed
by sustained effort on harder aspects. Overall, the learning effort framework
provides a tractable theoretical test bed to study normative benefits of
interventions in a variety of learning systems, as well as a formal account of
optimal cognitive control strategies over learning trajectories posited by
established theories in cognitive neuroscience.Comment: Under Revie
- …