964 research outputs found
European information: a guide to official sources
Fourth edition revised by Eric Davies. Originally published online by European Information AssociationA practical guide for anyone who needs to find official information about the European institutions and their activitie
Special Libraries, Winter 1986
Volume 77, Issue 1https://scholarworks.sjsu.edu/sla_sl_1986/1000/thumbnail.jp
Medical Informatics
Information technology has been revolutionizing the everyday life of the common man, while medical science has been making rapid strides in understanding disease mechanisms, developing diagnostic techniques and effecting successful treatment regimen, even for those cases which would have been classified as a poor prognosis a decade earlier. The confluence of information technology and biomedicine has brought into its ambit additional dimensions of computerized databases for patient conditions, revolutionizing the way health care and patient information is recorded, processed, interpreted and utilized for improving the quality of life. This book consists of seven chapters dealing with the three primary issues of medical information acquisition from a patient's and health care professional's perspective, translational approaches from a researcher's point of view, and finally the application potential as required by the clinicians/physician. The book covers modern issues in Information Technology, Bioinformatics Methods and Clinical Applications. The chapters describe the basic process of acquisition of information in a health system, recent technological developments in biomedicine and the realistic evaluation of medical informatics
Foundation, Implementation and Evaluation of the MorphoSaurus System: Subword Indexing, Lexical Learning and Word Sense Disambiguation for Medical Cross-Language Information Retrieval
Im medizinischen Alltag, zu welchem viel Dokumentations- und Recherchearbeit gehört, ist mittlerweile der überwiegende Teil textuell kodierter Information elektronisch verfügbar. Hiermit kommt der Entwicklung leistungsfähiger Methoden zur effizienten Recherche eine vorrangige Bedeutung zu.
Bewertet man die Nützlichkeit gängiger Textretrievalsysteme aus dem Blickwinkel der medizinischen Fachsprache, dann mangelt es ihnen an morphologischer Funktionalität (Flexion, Derivation und Komposition), lexikalisch-semantischer Funktionalität und der Fähigkeit zu einer sprachübergreifenden Analyse großer Dokumentenbestände.
In der vorliegenden Promotionsschrift werden die theoretischen Grundlagen des MorphoSaurus-Systems (ein Akronym für Morphem-Thesaurus) behandelt. Dessen methodischer Kern stellt ein um Morpheme der medizinischen Fach- und Laiensprache gruppierter Thesaurus dar, dessen Einträge mittels semantischer Relationen sprachübergreifend verknüpft sind. Darauf aufbauend wird ein Verfahren vorgestellt, welches (komplexe) Wörter in Morpheme segmentiert, die durch sprachunabhängige, konzeptklassenartige Symbole ersetzt werden. Die resultierende Repräsentation ist die Basis für das sprachübergreifende, morphemorientierte Textretrieval.
Neben der Kerntechnologie wird eine Methode zur automatischen Akquise von Lexikoneinträgen vorgestellt, wodurch bestehende Morphemlexika um weitere Sprachen ergänzt werden. Die Berücksichtigung sprachübergreifender Phänomene führt im Anschluss zu einem neuartigen Verfahren zur Auflösung von semantischen Ambiguitäten.
Die Leistungsfähigkeit des morphemorientierten Textretrievals wird im Rahmen umfangreicher, standardisierter Evaluationen empirisch getestet und gängigen Herangehensweisen gegenübergestellt
European information: a guide to official sources
Revised version of European information : a guide to official sources. Manchester, EIA, 2005.ISBN 1905106092. This edition published online only.This book is intended as a practical guide for anyone who needs to find official information, whether printed or
online, about the European institutions and their activities
Agile in-litero experiments:how can semi-automated information extraction from neuroscientific literature help neuroscience model building?
In neuroscience, as in many other scientific domains, the primary form of knowledge dissemination is through published articles in peer-reviewed journals. One challenge for modern neuroinformatics is to design methods to make the knowledge from the tremendous backlog of publications accessible for search, analysis and its integration into computational models. In this thesis, we introduce novel natural language processing (NLP) models and systems to mine the neuroscientific literature. In addition to in vivo, in vitro or in silico experiments, we coin the NLP methods developed in this thesis as in litero experiments, aiming at analyzing and making accessible the extended body of neuroscientific literature. In particular, we focus on two important neuroscientific entities: brain regions and neural cells. An integrated NLP model is designed to automatically extract brain region connectivity statements from very large corpora. This system is applied to a large corpus of 25M PubMed abstracts and 600K full-text articles. Central to this system is the creation of a searchable database of brain region connectivity statements, allowing neuroscientists to gain an overview of all brain regions connected to a given region of interest. More importantly, the database enables researcher to provide feedback on connectivity results and links back to the original article sentence to provide the relevant context. The database is evaluated by neuroanatomists on real connectomics tasks (targets of Nucleus Accumbens) and results in significant effort reduction in comparison to previous manual methods (from 1 week to 2h). Subsequently, we introduce neuroNER to identify, normalize and compare instances of identify neuronsneurons in the scientific literature. Our method relies on identifying and analyzing each of the domain features used to annotate a specific neuron mention, like the morphological term 'basket' or brain region 'hippocampus'. We apply our method to the same corpus of 25M PubMed abstracts and 600K full-text articles and find over 500K unique neuron type mentions. To demonstrate the utility of our approach, we also apply our method towards cross-comparing the NeuroLex and Human Brain Project (HBP) cell type ontologies. By decoupling a neuron mention's identity into its specific compositional features, our method can successfully identify specific neuron types even if they are not explicitly listed within a predefined neuron type lexicon, thus greatly facilitating cross-laboratory studies. In order to build such large databases, several tools and infrastructureslarge-scale NLP were developed: a robust pipeline to preprocess full-text PDF articles, as well as bluima, an NLP processing pipeline specialized on neuroscience to perform text-mining at PubMed scale. During the development of those two NLP systems, we acknowledged the need for novel NLP approaches to rapidly develop custom text mining solutions. This led to the formalization of the agile text miningagile text-mining methodology to improve the communication and collaboration between subject matter experts and text miners. Agile text mining is characterized by short development cycles, frequent tasks redefinition and continuous performance monitoring through integration tests. To support our approach, we developed Sherlok, an NLP framework designed for the development of agile text mining applications
Towards a system of concepts for Family Medicine. Multilingual indexing in General Practice/ Family Medicine in the era of Semantic Web
UNIVERSITY OF LIÈGE, BELGIUM
Executive Summary
Faculty of Medicine
Département Universitaire de Médecine Générale.
Unité de recherche Soins Primaires et Santé
Doctor in biomedical sciences
Towards a system of concepts for Family Medicine.
Multilingual indexing in General Practice/ Family Medicine in the era
of SemanticWeb
by Dr. Marc JAMOULLE
Introduction
This thesis is about giving visibility to the often overlooked work of family
physicians and consequently, is about grey literature in General Practice
and Family Medicine (GP/FM). It often seems that conference organizers
do not think of GP/FM as a knowledge-producing discipline that deserves
active dissemination. A conference is organized, but not much is done with
the knowledge shared at these meetings. In turn, the knowledge cannot be
reused or reapplied. This these is also about indexing. To find knowledge
back, indexing is mandatory. We must prepare tools that will automatically
index the thousands of abstracts that family doctors produce each year in
various languages. And finally this work is about semantics1. It is an introduction
to health terminologies, ontologies, semantic data, and linked
open data. All are expressions of the next step: Semantic Web for health
care data. Concepts, units of thought expressed by terms, will be our target
and must have the ability to be expressed in multiple languages. In turn,
three areas of knowledge are at stake in this study: (i) Family Medicine as a
pillar of primary health care, (ii) computational linguistics, and (iii) health
information systems.
Aim
• To identify knowledge produced by General practitioners (GPs) by
improving annotation of grey literature in Primary Health Care
• To propose an experimental indexing system, acting as draft for a
standardized table of content of GP/GM
• To improve the searchability of repositories for grey literature in GP/GM.
1For specific terms, see the Glossary page 257
x
Methods
The first step aimed to design the taxonomy by identifying relevant concepts
in a compiled corpus of GP/FM texts. We have studied the concepts
identified in nearly two thousand communications of GPs during
conferences. The relevant concepts belong to the fields that are focusing
on GP/FM activities (e.g. teaching, ethics, management or environmental
hazard issues).
The second step was the development of an on-line, multilingual, terminological
resource for each category of the resulting taxonomy, named
Q-Codes. We have designed this terminology in the form of a lightweight
ontology, accessible on-line for readers and ready for use by computers of
the semantic web. It is also fit for the Linked Open Data universe.
Results
We propose 182 Q-Codes in an on-line multilingual database (10 languages)
(www.hetop.eu/Q) acting each as a filter for Medline. Q-Codes are also available
under the form of Unique Resource Identifiers (URIs) and are exportable
in Web Ontology Language (OWL). The International Classification of Primary
Care (ICPC) is linked to Q-Codes in order to form the Core Content
Classification in General Practice/Family Medicine (3CGP). So far, 3CGP is
in use by humans in pedagogy, in bibliographic studies, in indexing congresses,
master theses and other forms of grey literature in GP/FM. Use by
computers is experimented in automatic classifiers, annotators and natural
language processing.
Discussion
To the best of our knowledge, this is the first attempt to expand the ICPC
coding system with an extension for family physician contextual issues,
thus covering non-clinical content of practice. It remains to be proven that
our proposed terminology will help in dealing with more complex systems,
such as MeSH, to support information storage and retrieval activities.
However, this exercise is proposed as a first step in the creation of an ontology
of GP/FM and as an opening to the complex world of Semantic Web
technologies.
Conclusion
We expect that the creation of this terminological resource for indexing abstracts
and for facilitating Medline searches for general practitioners, researchers
and students in medicine will reduce loss of knowledge in the
domain of GP/FM. In addition, through better indexing of the grey literature
(congress abstracts, master’s and doctoral theses), we hope to enhance
the accessibility of research results and give visibility to the invisible work
of family physicians
Internet technologies relevant to private investigators’ working practices
Much has been written and discussed especially in the various US media and in legislative organs, about how the Internet is used illegally (hacking, stalking for instance), but hardly if any research has been done as to how the investigative industry employs the new medium to its benefit. The author described in this thesis how private investigators (PIs) execute their profession these days using the facilities the Internet avail them in contrast to the time before the dawn of the Internet. This contrast is also investigated in an international context, an important part of the thesis, drawn from the author's 32 years of international business experience and that of PIs worldwide. The availability of the various online facilities in different countries are compared. To better understand the new medium, and its facilities a short outline of the Internet’s history, it’s set up in general and for the use of PIs in particular is supplied. PIs also face limitations in their daily work, limitations originating from online, legal, educational, financial and international causes. The new medium not only helps PIs in their investigative, but also in their office work. Finally PIs' wishes for new tools to facilitate their daily investigative work and their outlook as to where the new medium will head are also discussed
Simple identification tools in FishBase
Simple identification tools for fish species were included in the FishBase information system from its inception. Early tools made use of the relational model and characters like fin ray meristics. Soon pictures and drawings were added as a further help, similar to a field guide. Later came the computerization of existing dichotomous keys, again in combination with pictures and other information, and the ability to restrict possible species by country, area, or taxonomic group. Today, www.FishBase.org offers four different ways to identify species. This paper describes these tools with their advantages and disadvantages, and suggests various options for further
development. It explores the possibility of a holistic and integrated computeraided strategy
- …