4,804 research outputs found
Scheduling with partial orders and a causal model
In an ongoing project at Honeywell SRC, we are constructing a prototype scheduling system for a NASA domain using the 'Time Map Manager' (TMM). The TMM representations are flexible enough to permit the representation of precedence constraints, metric constraints between activities, and constraints relative to a variety of references (e.g., Mission Elapsed Time vs. Mission Day). The TMM also supports a simple form of causal reasoning (projection), dynamic database updates, and monitoring specified database properties as changes occur over time. The greatest apparent advantage to using the TMM is the flexibility added to the scheduling process: schedules are constructed by a process of 'iterative refinement,' in which scheduling decisions correspond to constraining an activity either with respect to another activity or with respect to one time line. The schedule becomes more detailed as activities and constraints are added. Undoing a scheduling decision means removing a constraint, not removing an activity from a specified place on the time line. For example, we can move an activity around on the time line by deleting constraints and adding new ones
Recommended from our members
Deep learning for cardiac image segmentation: A review
Deep learning has become the most widely used approach for cardiac image segmentation in recent years. In this paper, we provide a review of over 100 cardiac image segmentation papers using deep learning, which covers common imaging modalities including magnetic resonance imaging (MRI), computed tomography (CT), and ultrasound (US) and major anatomical structures of interest (ventricles, atria and vessels). In addition, a summary of publicly available cardiac image datasets and code repositories are included to provide a base for encouraging reproducible research. Finally, we discuss the challenges and limitations with current deep learning-based approaches (scarcity of labels, model generalizability across different domains, interpretability) and suggest potential directions for future research
Towards Automated Urban Planning: When Generative and ChatGPT-like AI Meets Urban Planning
The two fields of urban planning and artificial intelligence (AI) arose and
developed separately. However, there is now cross-pollination and increasing
interest in both fields to benefit from the advances of the other. In the
present paper, we introduce the importance of urban planning from the
sustainability, living, economic, disaster, and environmental perspectives. We
review the fundamental concepts of urban planning and relate these concepts to
crucial open problems of machine learning, including adversarial learning,
generative neural networks, deep encoder-decoder networks, conversational AI,
and geospatial and temporal machine learning, thereby assaying how AI can
contribute to modern urban planning. Thus, a central problem is automated
land-use configuration, which is formulated as the generation of land uses and
building configuration for a target area from surrounding geospatial, human
mobility, social media, environment, and economic activities. Finally, we
delineate some implications of AI for urban planning and propose key research
areas at the intersection of both topics.Comment: TSAS Submissio
Bayesian Active Learning for Personalization and Uncertainty Quantification in Cardiac Electrophysiological Model
Cardiacvascular disease is the top death causing disease worldwide. In recent years, high-fidelity personalized models of the heart have shown an increasing capability to supplement clinical cardiology for improved patient-specific diagnosis, prediction, and treatment planning. In addition, they have shown promise to improve scientific understanding of a variety of disease mechanisms.
However, model personalization by estimating the patient-specific tissue properties that are in the form of parameters of a physiological model is challenging. This is because tissue properties, in general, cannot be directly measured and they need to be estimated from measurements that are indirectly related to them through a physiological model. Moreover, these unknown tissue properties are heterogeneous and spatially varying throughout the heart volume presenting a difficulty of high-dimensional (HD) estimation from indirect and limited measurement data. The challenge in model personalization, therefore, summarizes to solving an ill-posed inverse problem where the unknown parameters are HD and the forward model is complex with a non-linear and computationally expensive physiological model.
In this dissertation, we address the above challenge with following contributions. First, to address the concern of a complex forward model, we propose the surrogate modeling of the complex target function containing the forward model – an objective function in deterministic estimation or a posterior probability density function in probabilistic estimation – by actively selecting a set of training samples and a Bayesian update of the prior over the target function. The efficient and accurate surrogate of the expensive target function obtained in this manner is then utilized to accelerate either deterministic or probabilistic parameter estimation. Next, within the framework of Bayesian active learning we enable active surrogate learning over a HD parameter space with two novel approaches: 1) a multi-scale optimization that can adaptively allocate higher resolution to heterogeneous tissue regions and lower resolution to homogeneous tissue regions; and 2) a generative model from low-dimensional (LD) latent code to HD tissue properties. Both of these approaches are independently developed and tested within a parameter optimization framework. Furthermore, we devise a novel method that utilizes the surrogate pdf learned on an estimated LD parameter space to improve the proposal distribution of Metropolis Hastings for an accelerated sampling of the exact posterior pdf. We evaluate the presented methods on estimating local tissue excitability of a cardiac electrophysiological model in both synthetic data experiments and real data experiments. Results demonstrate that the presented methods are able to improve the accuracy and efficiency in patient-specific model parameter estimation in comparison to the existing approaches used for model personalization
Clause-Type, Force, and Normative Judgment in the Semantics of Imperatives
I argue that imperatives express contents that are both cognitively and semantically related to, but nevertheless distinct from, modal propositions. Imperatives, on this analysis, semantically encode features of planning that are modally specified. Uttering an imperative amounts to tokening this feature in discourse, and thereby proffering it for adoption by the audience. This analysis deals smoothly with the problems afflicting Portner's Dynamic Pragmatic account and Kaufmann's Modal account. It also suggests an appealing reorientation of clause-type theorizing, in which the cognitive act of updating on a typed sentence plays a central role in theorizing about both its semantics and role in discourse
- …