11,825 research outputs found
Large-Scale Analysis of the Accuracy of the Journal Classification Systems of Web of Science and Scopus
Journal classification systems play an important role in bibliometric
analyses. The two most important bibliographic databases, Web of Science and
Scopus, each provide a journal classification system. However, no study has
systematically investigated the accuracy of these classification systems. To
examine and compare the accuracy of journal classification systems, we define
two criteria on the basis of direct citation relations between journals and
categories. We use Criterion I to select journals that have weak connections
with their assigned categories, and we use Criterion II to identify journals
that are not assigned to categories with which they have strong connections. If
a journal satisfies either of the two criteria, we conclude that its assignment
to categories may be questionable. Accordingly, we identify all journals with
questionable classifications in Web of Science and Scopus. Furthermore, we
perform a more in-depth analysis for the field of Library and Information
Science to assess whether our proposed criteria are appropriate and whether
they yield meaningful results. It turns out that according to our
citation-based criteria Web of Science performs significantly better than
Scopus in terms of the accuracy of its journal classification system
A review of the literature on citation impact indicators
Citation impact indicators nowadays play an important role in research
evaluation, and consequently these indicators have received a lot of attention
in the bibliometric and scientometric literature. This paper provides an
in-depth review of the literature on citation impact indicators. First, an
overview is given of the literature on bibliographic databases that can be used
to calculate citation impact indicators (Web of Science, Scopus, and Google
Scholar). Next, selected topics in the literature on citation impact indicators
are reviewed in detail. The first topic is the selection of publications and
citations to be included in the calculation of citation impact indicators. The
second topic is the normalization of citation impact indicators, in particular
normalization for field differences. Counting methods for dealing with
co-authored publications are the third topic, and citation impact indicators
for journals are the last topic. The paper concludes by offering some
recommendations for future research
Construction of a Pragmatic Base Line for Journal Classifications and Maps Based on Aggregated Journal-Journal Citation Relations
A number of journal classification systems have been developed in
bibliometrics since the launch of the Citation Indices by the Institute of
Scientific Information (ISI) in the 1960s. These systems are used to normalize
citation counts with respect to field-specific citation patterns. The best
known system is the so-called "Web-of-Science Subject Categories" (WCs). In
other systems papers are classified by algorithmic solutions. Using the Journal
Citation Reports 2014 of the Science Citation Index and the Social Science
Citation Index (n of journals = 11,149), we examine options for developing a
new system based on journal classifications into subject categories using
aggregated journal-journal citation data. Combining routines in VOSviewer and
Pajek, a tree-like classification is developed. At each level one can generate
a map of science for all the journals subsumed under a category. Nine major
fields are distinguished at the top level. Further decomposition of the social
sciences is pursued for the sake of example with a focus on journals in
information science (LIS) and science studies (STS). The new classification
system improves on alternative options by avoiding the problem of randomness in
each run that has made algorithmic solutions hitherto irreproducible.
Limitations of the new system are discussed (e.g. the classification of
multi-disciplinary journals). The system's usefulness for field-normalization
in bibliometrics should be explored in future studies.Comment: accepted for publication in the Journal of Informetrics, 20 July 201
"Seed+Expand": A validated methodology for creating high quality publication oeuvres of individual researchers
The study of science at the individual micro-level frequently requires the
disambiguation of author names. The creation of author's publication oeuvres
involves matching the list of unique author names to names used in publication
databases. Despite recent progress in the development of unique author
identifiers, e.g., ORCID, VIVO, or DAI, author disambiguation remains a key
problem when it comes to large-scale bibliometric analysis using data from
multiple databases. This study introduces and validates a new methodology
called seed+expand for semi-automatic bibliographic data collection for a given
set of individual authors. Specifically, we identify the oeuvre of a set of
Dutch full professors during the period 1980-2011. In particular, we combine
author records from the National Research Information System (NARCIS) with
publication records from the Web of Science. Starting with an initial list of
8,378 names, we identify "seed publications" for each author using five
different approaches. Subsequently, we "expand" the set of publication in three
different approaches. The different approaches are compared and resulting
oeuvres are evaluated on precision and recall using a "gold standard" dataset
of authors for which verified publications in the period 2001-2010 are
available.Comment: Paper accepted for the ISSI 2013, small changes in the text due to
referee comments, one figure added (Fig 3
The Research Space: using the career paths of scholars to predict the evolution of the research output of individuals, institutions, and nations
In recent years scholars have built maps of science by connecting the
academic fields that cite each other, are cited together, or that cite a
similar literature. But since scholars cannot always publish in the fields they
cite, or that cite them, these science maps are only rough proxies for the
potential of a scholar, organization, or country, to enter a new academic
field. Here we use a large dataset of scholarly publications disambiguated at
the individual level to create a map of science-or research space-where links
connect pairs of fields based on the probability that an individual has
published in both of them. We find that the research space is a significantly
more accurate predictor of the fields that individuals and organizations will
enter in the future than citation based science maps. At the country level,
however, the research space and citations based science maps are equally
accurate. These findings show that data on career trajectories-the set of
fields that individuals have previously published in-provide more accurate
predictors of future research output for more focalized units-such as
individuals or organizations-than citation based science maps
The scholarly impact of TRECVid (2003-2009)
This paper reports on an investigation into the scholarly impact of the TRECVid (TREC Video Retrieval Evaluation) benchmarking conferences between 2003 and 2009. The contribution of TRECVid to research in video retrieval is assessed by analyzing publication content to show the development of techniques and approaches over time and by analyzing publication impact through publication numbers and citation analysis. Popular conference and journal venues for TRECVid publications are identified in terms of number of citations received. For a selection of participants at different career stages, the relative importance of TRECVid publications in terms of citations vis a vis their other publications is investigated. TRECVid, as an evaluation conference, provides data on which research teams âscoredâ highly against the evaluation criteria and the relationship between âtop scoringâ teams at TRECVid and the âtop scoringâ papers in terms of citations is analysed. A strong relationship was found between âsuccessâ at TRECVid and âsuccessâ at citations both for high scoring and low scoring teams. The implications of the study in terms of the value of TRECVid as a research activity, and the value of bibliometric analysis as a research evaluation tool, are discussed
- âŠ