40 research outputs found

    A Novel Graph Constructor for Semisupervised Discriminant Analysis: Combined Low-Rank and k

    Get PDF
    Semisupervised Discriminant Analysis (SDA) is a semisupervised dimensionality reduction algorithm, which can easily resolve the out-of-sample problem. Relative works usually focus on the geometric relationships of data points, which are not obvious, to enhance the performance of SDA. Different from these relative works, the regularized graph construction is researched here, which is important in the graph-based semisupervised learning methods. In this paper, we propose a novel graph for Semisupervised Discriminant Analysis, which is called combined low-rank and k-nearest neighbor (LRKNN) graph. In our LRKNN graph, we map the data to the LR feature space and then the kNN is adopted to satisfy the algorithmic requirements of SDA. Since the low-rank representation can capture the global structure and the k-nearest neighbor algorithm can maximally preserve the local geometrical structure of the data, the LRKNN graph can significantly improve the performance of SDA. Extensive experiments on several real-world databases show that the proposed LRKNN graph is an efficient graph constructor, which can largely outperform other commonly used baselines

    Improving Emergency Department Patient Flow Through Near Real-Time Analytics

    Get PDF
    ABSTRACT IMPROVING EMERGENCY DEPARTMENT PATIENT FLOW THROUGH NEAR REAL-TIME ANALYTICS This dissertation research investigates opportunities for developing effective decision support models that exploit near real-time (NRT) information to enhance the operational intelligence within hospital Emergency Departments (ED). Approaching from a systems engineering perspective, the study proposes a novel decision support framework for streamlining ED patient flow that employs machine learning, statistical and operations research methods to facilitate its operationalization. ED crowding has become the subject of significant public and academic attention, and it is known to cause a number of adverse outcomes to the patients, ED staff as well as hospital revenues. Despite many efforts to investigate the causes, consequences and interventions for ED overcrowding in the past two decades, scientific knowledge remains limited in regards to strategies and pragmatic approaches that actually improve patient flow in EDs. Motivated by the gaps in research, we develop a near real-time triage decision support system to reduce ED boarding and improve ED patient flow. The proposed system is a novel variant of a newsvendor modeling framework that integrates patient admission probability prediction within a proactive ward-bed reservation system to improve the effectiveness of bed coordination efforts and reduce boarding times for ED patients along with the resulting costs. Specifically, we propose a cost-sensitive bed reservation policy that recommends optimal bed reservation times for patients right during triage. The policy relies on classifiers that estimate the probability that the ED patient will be admitted using the patient information collected and readily available at triage or right after. The policy is cost-sensitive in that it accounts for costs associated with patient admission prediction misclassification as well as costs associated with incorrectly selecting the reservation time. To achieve the objective of this work, we also addressed two secondary objectives: first, development of models to predict the admission likelihood and target admission wards of ED patients; second, development of models to estimate length-of-stay (LOS) of ED patients. For the first secondary objective, we develop an algorithm that incorporates feature selection into a state-of-the-art and powerful probabilistic Bayesian classification method: multi-class relevance vector machine. For the second objective, we investigated the performance of hazard rate models (in particual, the non-parametric Cox proportional hazard model, parametric hazard rate models, as well as artificial neural networks for modeling the hazard rate) to estimate ED LOS by using the information that is available at triage or right after as the covariates in the models. The proposed models are tested using extensive historical data from several U.S. Department of Veterans Affairs Medical Centers (VAMCs) in the Mid-West. The Case Study using historical data from a VAMC demonstrates that applying the proposed framework leads to significant savings associated with reduced boarding times, in particular, for smaller wards with high levels of utilization. For theory, our primary contribution is the development of a cost sensitive ward-bed reservation model that effectively accounts for various costs and uncertainties. This work also contributes to the development of an integrated feature selection method for classification by developing and validating the mathematical derivation for feature selection during mRVM learning. Another contribution stems from investigating how much the ED LOS estimation can be improved by incorporating the information regarding ED orderable item lists. Overall, this work is a successful application of mixed methods of operation research, machine learning and statistics to the important domain of health care system efficiency improvement

    Cross-domain Recommendations based on semantically-enhanced User Web Behavior

    Get PDF
    Information seeking in the Web can be facilitated by recommender systems that guide the users in a personalized manner to relevant resources in the large space of the possible options in the Web. This work investigates how to model people\u27s Web behavior at multiple sites and learn to predict future preferences, in order to generate relevant cross-domain recommendations. This thesis contributes with novel techniques for building cross-domain recommender systems in an open Web setting

    Machine Learning

    Get PDF
    Machine Learning can be defined in various ways related to a scientific domain concerned with the design and development of theoretical and implementation tools that allow building systems with some Human Like intelligent behavior. Machine learning addresses more specifically the ability to improve automatically through experience

    Uncertainty in Artificial Intelligence: Proceedings of the Thirty-Fourth Conference

    Get PDF

    Application of Hierarchical Temporal Memory to Anomaly Detection of Vital Signs for Ambient Assisted Living

    Get PDF
    This thesis presents the development of a framework for anomaly detection of vital signs for an Ambient Assisted Living (AAL) health monitoring scenario. It is driven by spatiotemporal reasoning of vital signs that Cortical Learning Algorithms (CLA) based on Hierarchal Temporal Memory (HTM) theory undertakes in an AAL health monitoring scenario to detect anomalous data points preceding cardiac arrest. This thesis begins with a literature review on the existing Ambient intelligence (AmI) paradigm, AAL technologies and anomaly detection algorithms used in a health monitoring scenario. The research revealed the significance of the temporal and spatial reasoning in the vital signs monitoring as the spatiotemporal patterns of vital signs provide a basis to detect irregularities in the health status of elderly people. The HTM theory is yet to be adequately deployed in an AAL health monitoring scenario. Hence HTM theory, network and core operations of the CLA are explored. Despite the fact that standard implementation of the HTM theory comprises of a single-level hierarchy, multiple vital signs, specifically the correlation between them is not sufficiently considered. This insufficiency is of particular significance considering that vital signs are correlated in time and space, which are used in the health monitoring applications for diagnosis and prognosis tasks. This research proposes a novel framework consisting of multi-level HTM networks. The lower level consists of four models allocated to the four vital signs, Systolic Blood Pressure (SBP), Diastolic Blood Pressure (DBP), Heart Rate (HR) and peripheral capillary oxygen saturation (SpO2) in order to learn the spatiotemporal patterns of each vital sign. Additionally, a higher level is introduced to learn spatiotemporal patterns of the anomalous data point detected from the four vital signs. The proposed hierarchical organisation improves the model’s performance by using the semantically improved representation of the sensed data because patterns learned at each level of the hierarchy are reused when combined in novel ways at higher levels. To investigate and evaluate the performance of the proposed framework, several data selection techniques are studied, and accordingly, a total record of 247 elderly patients is extracted from the MIMIC-III clinical database. The performance of the proposed framework is evaluated and compared against several state-of-the-art anomaly detection algorithms using both online and traditional metrics. The proposed framework achieved 83% NAB score which outperforms the HTM and k-NN algorithms by 15%, the HBOS and INFLO SVD by 16% and the k-NN PCA by 21% while the SVM scored 34%. The results prove that multiple HTM networks can achieve better performance when dealing with multi-dimensional data, i.e. data collected from more than one source/sensor

    Semantic multimedia modelling & interpretation for annotation

    Get PDF
    The emergence of multimedia enabled devices, particularly the incorporation of cameras in mobile phones, and the accelerated revolutions in the low cost storage devices, boosts the multimedia data production rate drastically. Witnessing such an iniquitousness of digital images and videos, the research community has been projecting the issue of its significant utilization and management. Stored in monumental multimedia corpora, digital data need to be retrieved and organized in an intelligent way, leaning on the rich semantics involved. The utilization of these image and video collections demands proficient image and video annotation and retrieval techniques. Recently, the multimedia research community is progressively veering its emphasis to the personalization of these media. The main impediment in the image and video analysis is the semantic gap, which is the discrepancy among a user’s high-level interpretation of an image and the video and the low level computational interpretation of it. Content-based image and video annotation systems are remarkably susceptible to the semantic gap due to their reliance on low-level visual features for delineating semantically rich image and video contents. However, the fact is that the visual similarity is not semantic similarity, so there is a demand to break through this dilemma through an alternative way. The semantic gap can be narrowed by counting high-level and user-generated information in the annotation. High-level descriptions of images and or videos are more proficient of capturing the semantic meaning of multimedia content, but it is not always applicable to collect this information. It is commonly agreed that the problem of high level semantic annotation of multimedia is still far from being answered. This dissertation puts forward approaches for intelligent multimedia semantic extraction for high level annotation. This dissertation intends to bridge the gap between the visual features and semantics. It proposes a framework for annotation enhancement and refinement for the object/concept annotated images and videos datasets. The entire theme is to first purify the datasets from noisy keyword and then expand the concepts lexically and commonsensical to fill the vocabulary and lexical gap to achieve high level semantics for the corpus. This dissertation also explored a novel approach for high level semantic (HLS) propagation through the images corpora. The HLS propagation takes the advantages of the semantic intensity (SI), which is the concept dominancy factor in the image and annotation based semantic similarity of the images. As we are aware of the fact that the image is the combination of various concepts and among the list of concepts some of them are more dominant then the other, while semantic similarity of the images are based on the SI and concept semantic similarity among the pair of images. Moreover, the HLS exploits the clustering techniques to group similar images, where a single effort of the human experts to assign high level semantic to a randomly selected image and propagate to other images through clustering. The investigation has been made on the LabelMe image and LabelMe video dataset. Experiments exhibit that the proposed approaches perform a noticeable improvement towards bridging the semantic gap and reveal that our proposed system outperforms the traditional systems

    Social media mental health analysis framework through applied computational approaches

    Get PDF
    Studies have shown that mental illness burdens not only public health and productivity but also established market economies throughout the world. However, mental disorders are difficult to diagnose and monitor through traditional methods, which heavily rely on interviews, questionnaires and surveys, resulting in high under-diagnosis and under-treatment rates. The increasing use of online social media, such as Facebook and Twitter, is now a common part of people’s everyday life. The continuous and real-time user-generated content often reflects feelings, opinions, social status and behaviours of individuals, creating an unprecedented wealth of person-specific information. With advances in data science, social media has already been increasingly employed in population health monitoring and more recently mental health applications to understand mental disorders as well as to develop online screening and intervention tools. However, existing research efforts are still in their infancy, primarily aimed at highlighting the potential of employing social media in mental health research. The majority of work is developed on ad hoc datasets and lacks a systematic research pipeline. [Continues.]</div

    Q(sqrt(-3))-Integral Points on a Mordell Curve

    Get PDF
    We use an extension of quadratic Chabauty to number fields,recently developed by the author with Balakrishnan, Besser and M ̈uller,combined with a sieving technique, to determine the integral points overQ(√−3) on the Mordell curve y2 = x3 − 4
    corecore