37,766 research outputs found
Conceptual biology, hypothesis discovery, and text mining: Swanson's legacy
Innovative biomedical librarians and information specialists who want to expand their roles as expert searchers need to know about profound changes in biology and parallel trends in text mining. In recent years, conceptual biology has emerged as a complement to empirical biology. This is partly in response to the availability of massive digital resources such as the network of databases for molecular biologists at the National Center for Biotechnology Information. Developments in text mining and hypothesis discovery systems based on the early work of Swanson, a mathematician and information scientist, are coincident with the emergence of conceptual biology. Very little has been written to introduce biomedical digital librarians to these new trends. In this paper, background for data and text mining, as well as for knowledge discovery in databases (KDD) and in text (KDT) is presented, then a brief review of Swanson's ideas, followed by a discussion of recent approaches to hypothesis discovery and testing. 'Testing' in the context of text mining involves partially automated methods for finding evidence in the literature to support hypothetical relationships. Concluding remarks follow regarding (a) the limits of current strategies for evaluation of hypothesis discovery systems and (b) the role of literature-based discovery in concert with empirical research. Report of an informatics-driven literature review for biomarkers of systemic lupus erythematosus is mentioned. Swanson's vision of the hidden value in the literature of science and, by extension, in biomedical digital databases, is still remarkably generative for information scientists, biologists, and physicians. © 2006Bekhuis; licensee BioMed Central Ltd
Deep Learning based Recommender System: A Survey and New Perspectives
With the ever-growing volume of online information, recommender systems have
been an effective strategy to overcome such information overload. The utility
of recommender systems cannot be overstated, given its widespread adoption in
many web applications, along with its potential impact to ameliorate many
problems related to over-choice. In recent years, deep learning has garnered
considerable interest in many research fields such as computer vision and
natural language processing, owing not only to stellar performance but also the
attractive property of learning feature representations from scratch. The
influence of deep learning is also pervasive, recently demonstrating its
effectiveness when applied to information retrieval and recommender systems
research. Evidently, the field of deep learning in recommender system is
flourishing. This article aims to provide a comprehensive review of recent
research efforts on deep learning based recommender systems. More concretely,
we provide and devise a taxonomy of deep learning based recommendation models,
along with providing a comprehensive summary of the state-of-the-art. Finally,
we expand on current trends and provide new perspectives pertaining to this new
exciting development of the field.Comment: The paper has been accepted by ACM Computing Surveys.
https://doi.acm.org/10.1145/328502
AutoSVD++: An Efficient Hybrid Collaborative Filtering Model via Contractive Auto-encoders
Collaborative filtering (CF) has been successfully used to provide users with
personalized products and services. However, dealing with the increasing
sparseness of user-item matrix still remains a challenge. To tackle such issue,
hybrid CF such as combining with content based filtering and leveraging side
information of users and items has been extensively studied to enhance
performance. However, most of these approaches depend on hand-crafted feature
engineering, which are usually noise-prone and biased by different feature
extraction and selection schemes. In this paper, we propose a new hybrid model
by generalizing contractive auto-encoder paradigm into matrix factorization
framework with good scalability and computational efficiency, which jointly
model content information as representations of effectiveness and compactness,
and leverage implicit user feedback to make accurate recommendations. Extensive
experiments conducted over three large scale real datasets indicate the
proposed approach outperforms the compared methods for item recommendation.Comment: 4 pages, 3 figure
Assessing Perceptions of the Integrative Justice Model Propositions: A Critical Step Toward Operationalizing a Macro Model
Propositions underlying the theoretical tenets of the Integrative Justice Model (IJM), a normative, ethical framework for engaging in impoverished markets, are investigated for reliability in application. Santos and Laczniak provide numerous decision principles to help the marketer to evaluate fairness in the marketplace exchange, yet which are most reflective of the core tenets and most useful in application? Managerial perceptions of the extent to which the propositions reflect the core tenets of the IJM are evaluated in this critical step toward operationalizing the model. Factor analysis is implemented with a relatively small sample, a challenge particularly common in research with the impoverished or marginalized, to evaluate the decision principles most reflective of the core tenets, from the perspective of the marketer/manager. Normality is not considered a critical assumption of factor analysis where groups of like variables are clustered into underlying constructs. This work articulates a critical step in macromarketing research methods, exemplifying an approach resilient to micro samples under macro frameworks
The systematic guideline review: method, rationale, and test on chronic heart failure
Background: Evidence-based guidelines have the potential to improve healthcare. However, their de-novo-development requires substantial resources-especially for complex conditions, and adaptation may be biased by contextually influenced recommendations in source guidelines. In this paper we describe a new approach to guideline development-the systematic guideline review method (SGR), and its application in the development of an evidence-based guideline for family physicians on chronic heart failure (CHF).
Methods: A systematic search for guidelines was carried out. Evidence-based guidelines on CHF management in adults in ambulatory care published in English or German between the years 2000 and 2004 were included. Guidelines on acute or right heart failure were excluded. Eligibility was assessed by two reviewers, methodological quality of selected guidelines was appraised using the AGREE instrument, and a framework of relevant clinical questions for diagnostics and treatment was derived. Data were extracted into evidence tables, systematically compared by means of a consistency analysis and synthesized in a preliminary draft. Most relevant primary sources were re-assessed to verify the cited evidence. Evidence and recommendations were summarized in a draft guideline.
Results: Of 16 included guidelines five were of good quality. A total of 35 recommendations were systematically compared: 25/35 were consistent, 9/35 inconsistent, and 1/35 un-rateable (derived from a single guideline). Of the 25 consistencies, 14 were based on consensus, seven on evidence and four differed in grading. Major inconsistencies were found in 3/9 of the inconsistent recommendations. We re-evaluated the evidence for 17 recommendations (evidence-based, differing evidence levels and minor inconsistencies) - the majority was congruent. Incongruity was found where the stated evidence could not be verified in the cited primary sources, or where the evaluation in the source guidelines focused on treatment benefits and underestimated the risks. The draft guideline was completed in 8.5 man-months. The main limitation to this study was the lack of a second reviewer.
Conclusion: The systematic guideline review including framework development, consistency analysis and validation is an effective, valid, and resource saving-approach to the development of evidence-based guidelines
An Algorithmic Framework for Efficient Large-Scale Circuit Simulation Using Exponential Integrators
We propose an efficient algorithmic framework for time domain circuit
simulation using exponential integrator. This work addresses several critical
issues exposed by previous matrix exponential based circuit simulation
research, and makes it capable of simulating stiff nonlinear circuit system at
a large scale. In this framework, the system's nonlinearity is treated with
exponential Rosenbrock-Euler formulation. The matrix exponential and vector
product is computed using invert Krylov subspace method. Our proposed method
has several distinguished advantages over conventional formulations (e.g., the
well-known backward Euler with Newton-Raphson method). The matrix factorization
is performed only for the conductance/resistance matrix G, without being
performed for the combinations of the capacitance/inductance matrix C and
matrix G, which are used in traditional implicit formulations. Furthermore, due
to the explicit nature of our formulation, we do not need to repeat LU
decompositions when adjusting the length of time steps for error controls. Our
algorithm is better suited to solving tightly coupled post-layout circuits in
the pursuit for full-chip simulation. Our experimental results validate the
advantages of our framework.Comment: 6 pages; ACM/IEEE DAC 201
Text Analytics for Android Project
Most advanced text analytics and text mining tasks include text classification, text clustering, building ontology, concept/entity extraction, summarization, deriving patterns within the structured data, production of granular taxonomies, sentiment and emotion analysis, document summarization, entity relation modelling, interpretation of the output. Already existing text analytics and text mining cannot develop text material alternatives (perform a multivariant design), perform multiple criteria analysis,
automatically select the most effective variant according to different aspects (citation index of papers (Scopus, ScienceDirect, Google Scholar) and authors (Scopus, ScienceDirect, Google Scholar), Top 25 papers, impact factor of journals, supporting phrases, document name and contents, density of keywords), calculate utility degree and market value. However, the Text Analytics for Android Project can perform the aforementioned functions. To the best of the knowledge herein, these functions have not been previously implemented; thus this is the first attempt to do so. The Text Analytics for Android Project is briefly described in this article
Implicit reference to citations: a study of astronomy
The research in this paper presents results in the automatic classification of pronouns within articles into those which refer to cited research and those which do not. It also discusses the automatic linking of pronouns which do refer to citations to their corresponding citations. The current study focused on the pronoun they as used in papers in Astronomy journals. The paper describes a classifier trained on maximum entropy principles using features defined by the distance to preceding citations and the category of verbs associated to the pronoun under consideration
- …