578 research outputs found
Mapping the structure of science through clustering in citation networks : granularity, labeling and visualization
The science system is large, and millions of research publications are published each year.
Within the field of scientometrics, the features and characteristics of this system are studied
using quantitative methods. Research publications constitute a rich source of information
about the science system and a means to model and study science on a large scale. The
classification of research publications into fields is essential to answer many questions about
the features and characteristics of the science system.
Comprehensive, hierarchical, and detailed classifications of large sets of research publications
are not easy to obtain. A solution for this problem is to use network-based approaches to
cluster research publications based on their citation relations. Clustering approaches have
been applied to large sets of publications at the level of individual articles (in contrast to the
journal level) for about a decade. Such approaches are addressed in this thesis. I call the
resulting classifications “algorithmically constructed, publications-level classifications of
research publications” (ACPLCs).
The aim of the thesis is to improve interpretability and utility of ACPLCs. I focus on some
issues that hitherto have not received much attention in the previous literature: (1) Conceptual
framework. Such a framework is elaborated throughout the thesis. Using the social science
citation theory, I argue that citations contextualize and position publications in the science
system. Citations may therefore be used to identify research fields, defined as focus areas of
research at various granularity levels. (2) Granularity levels corresponding to conceptual
framework. In Articles I and II, a method is proposed on how to adjust the granularity of
ACPLCs in order to obtain clusters corresponding to research fields at two granularity levels:
topics and specialties. (3) Cluster labeling. Article III addresses labeling of clusters at
different semantic levels, from broad and large to narrow and small, and compares the use of
data from various bibliographic fields and different term weighting approaches. (4)
Visualization. The methods resulting from Articles I-III are applied in Article IV to obtain a
classification of about 19 million biomedical articles. I propose a visualization methodology
that provides overview of the classification, using clusters at coarse levels, as well as the
possibility to zoom into details, using clusters at a granular level.
In conclusion, I have improved interpretability and utility of ACPLCs by providing a
conceptual framework, adjusting granularity of clusters, labeling clusters and, finally, by
visualizing an ACPLC in a way that provides both overview and detail. I have demonstrated
how these methods can be applied to obtain ACPLCs that are useful to, for example, identify
and explore focus areas of research
Human-in-the-Loop Learning From Crowdsourcing and Social Media
Computational social studies using public social media data have become more and more popular because of the large amount of user-generated data available. The richness of social media data, coupled with noise and subjectivity, raise significant challenges for computationally studying social issues in a feasible and scalable manner. Machine learning problems are, as a result, often subjective or ambiguous when humans are involved. That is, humans solving the same problems might come to legitimate but completely different conclusions, based on their personal experiences and beliefs. When building supervised learning models, particularly when using crowdsourced training data, multiple annotations per data item are usually reduced to a single label representing ground truth. This inevitably hides a rich source of diversity and subjectivity of opinions about the labels.
Label distribution learning associates for each data item a probability distribution over the labels for that item, thus it can preserve diversities of opinions, beliefs, etc. that conventional learning hides or ignores. We propose a humans-in-the-loop learning framework to model and study large volumes of unlabeled subjective social media data with less human effort. We study various annotation tasks given to crowdsourced annotators and methods for aggregating their contributions in a manner that preserves subjectivity and disagreement. We introduce a strategy for learning label distributions with only five-to-ten labels per item by aggregating human-annotated labels over multiple, semantically related data items. We conduct experiments using our learning framework on data related to two subjective social issues (work and employment, and suicide prevention) that touch many people worldwide. Our methods can be applied to a broad variety of problems, particularly social problems. Our experimental results suggest that specific label aggregation methods can help provide reliable representative semantics at the population level
Foundation Metrics: Quantifying Effectiveness of Healthcare Conversations powered by Generative AI
Generative Artificial Intelligence is set to revolutionize healthcare
delivery by transforming traditional patient care into a more personalized,
efficient, and proactive process. Chatbots, serving as interactive
conversational models, will probably drive this patient-centered transformation
in healthcare. Through the provision of various services, including diagnosis,
personalized lifestyle recommendations, and mental health support, the
objective is to substantially augment patient health outcomes, all the while
mitigating the workload burden on healthcare providers. The life-critical
nature of healthcare applications necessitates establishing a unified and
comprehensive set of evaluation metrics for conversational models. Existing
evaluation metrics proposed for various generic large language models (LLMs)
demonstrate a lack of comprehension regarding medical and health concepts and
their significance in promoting patients' well-being. Moreover, these metrics
neglect pivotal user-centered aspects, including trust-building, ethics,
personalization, empathy, user comprehension, and emotional support. The
purpose of this paper is to explore state-of-the-art LLM-based evaluation
metrics that are specifically applicable to the assessment of interactive
conversational models in healthcare. Subsequently, we present an comprehensive
set of evaluation metrics designed to thoroughly assess the performance of
healthcare chatbots from an end-user perspective. These metrics encompass an
evaluation of language processing abilities, impact on real-world clinical
tasks, and effectiveness in user-interactive conversations. Finally, we engage
in a discussion concerning the challenges associated with defining and
implementing these metrics, with particular emphasis on confounding factors
such as the target audience, evaluation methods, and prompt techniques involved
in the evaluation process.Comment: 13 pages, 4 figures, 2 tables, journal pape
Living analytics methods for the social web
[no abstract
Term-driven E-Commerce
Die Arbeit nimmt sich der textuellen Dimension des E-Commerce an. Grundlegende Hypothese ist die textuelle Gebundenheit von Information und Transaktion im Bereich des elektronischen Handels. Überall dort, wo Produkte und Dienstleistungen angeboten, nachgefragt, wahrgenommen und bewertet werden, kommen natürlichsprachige Ausdrücke zum Einsatz. Daraus resultiert ist zum einen, wie bedeutsam es ist, die Varianz textueller Beschreibungen im E-Commerce zu erfassen, zum anderen können die umfangreichen textuellen Ressourcen, die bei E-Commerce-Interaktionen anfallen, im Hinblick auf ein besseres Verständnis natürlicher Sprache herangezogen werden
Automatic Generation of Personalized Recommendations in eCoaching
Denne avhandlingen omhandler eCoaching for personlig livsstilsstøtte i sanntid ved bruk av informasjons- og kommunikasjonsteknologi. Utfordringen er å designe, utvikle og teknisk evaluere en prototyp av en intelligent eCoach som automatisk genererer personlige og evidensbaserte anbefalinger til en bedre livsstil. Den utviklede løsningen er fokusert på forbedring av fysisk aktivitet. Prototypen bruker bærbare medisinske aktivitetssensorer. De innsamlede data blir semantisk representert og kunstig intelligente algoritmer genererer automatisk meningsfulle, personlige og kontekstbaserte anbefalinger for mindre stillesittende tid. Oppgaven bruker den veletablerte designvitenskapelige forskningsmetodikken for å utvikle teoretiske grunnlag og praktiske implementeringer. Samlet sett fokuserer denne forskningen på teknologisk verifisering snarere enn klinisk evaluering.publishedVersio
Mining a Small Medical Data Set by Integrating the Decision Tree and t-test
[[abstract]]Although several researchers have used statistical methods to prove that aspiration followed by the injection of 95% ethanol left in situ (retention) is an effective treatment for ovarian endometriomas, very few discuss the different conditions that could generate different recovery rates for the patients. Therefore, this study adopts the statistical method and decision tree techniques together to analyze the postoperative status of ovarian endometriosis patients under different conditions. Since our collected data set is small, containing only 212 records, we use all of these data as the training data. Therefore, instead of using a resultant tree to generate rules directly, we use the value of each node as a cut point to generate all possible rules from the tree first. Then, using t-test, we verify the rules to discover some useful description rules after all possible rules from the tree have been generated. Experimental results show that our approach can find some new interesting knowledge about recurrent ovarian endometriomas under different conditions.[[journaltype]]國外[[incitationindex]]EI[[booktype]]紙本[[countrycodes]]FI
- …