77 research outputs found
Crowd-Sourcing Fuzzy and Faceted Classification for Concept Search
Searching for concepts in science and technology is often a difficult task.
To facilitate concept search, different types of human-generated metadata have
been created to define the content of scientific and technical disclosures.
Classification schemes such as the International Patent Classification (IPC)
and MEDLINE's MeSH are structured and controlled, but require trained experts
and central management to restrict ambiguity (Mork, 2013). While unstructured
tags of folksonomies can be processed to produce a degree of structure
(Kalendar, 2010; Karampinas, 2012; Sarasua, 2012; Bragg, 2013) the freedom
enjoyed by the crowd typically results in less precision (Stock 2007).
Existing classification schemes suffer from inflexibility and ambiguity.
Since humans understand language, inference, implication, abstraction and hence
concepts better than computers, we propose to harness the collective wisdom of
the crowd. To do so, we propose a novel classification scheme that is
sufficiently intuitive for the crowd to use, yet powerful enough to facilitate
search by analogy, and flexible enough to deal with ambiguity. The system will
enhance existing classification information. Linking up with the semantic web
and computer intelligence, a Citizen Science effort (Good, 2013) would support
innovation by improving the quality of granted patents, reducing duplicitous
research, and stimulating problem-oriented solution design.
A prototype of our design is in preparation. A crowd-sourced fuzzy and
faceted classification scheme will allow for better concept search and improved
access to prior art in science and technology
Durham Zoo: powering a search-&-innovation engine with collective intelligence
Purpose – Durham Zoo (hereinafter – DZ) is a project to design and operate a concept search engine for science and technology. In DZ, a concept includes a solution to a problem in a particular context. Design – Concept searching is rendered complex by the fuzzy nature of a concept, the many possible implementations of the same concept, and the many more ways that the many implementations can be expressed in natural language. An additional complexity is the diversity of languages and formats, in which the concepts can be disclosed. Humans understand language, inference, implication and abstraction and, hence, concepts much better than computers, that in turn are much better at storing and processing vast amounts of data. We are 7 billion on the planet and we have the Internet as the backbone for Collective Intelligence. So, our concept search engine uses humans to store concepts via a shorthand that can be stored, processed and searched by computers: so, humans IN and computers OUT. The shorthand is classification: metadata in a structure that can define the content of a disclosure. The classification is designed to be powerful in terms of defining and searching concepts, whilst suited to a crowdsourcing effort. It is simple and intuitive to use. Most importantly, it is adapted to restrict ambiguity, which is the poison of classification, without imposing a restrictive centralised management. In the classification scheme, each entity is shown together in a graphical representation with related entities. The entities are arranged on a sliding scale of similarity. This sliding scale is effectively fuzzy classification. Findings – The authors of the paper have been developing a first classification scheme for the technology of traffic cones, this in preparation for a trial of a working system. The process has enabled the authors to further explore the practicalities of concept classification. The CmapTools knowledge modelling kit to develop the graphical representations has been used. Practical implications – Concept searching is seen as having two categories: prior art searching, which is searching for what already exists, and solution searching: a search for a
novel solution to an existing problem. Prior art searching is not as efficient a process, as all encompassing in scope, or as
accurate in result, as it could and probably should be. The prior art includes library collections, journals, conference proceedings and everything else that has been written, drawn, spoken or made public in any way. Much technical information is only published in patents. There is a good reason to improve prior art searching: research, industry, and
indeed humanity faces the spectre of patent thickets: an impenetrable legal space that effectively hinders innovation rather than promotes it. Improved prior-art searching would help with the gardening and result in fewer and higher-quality patents. Poor-quality patents can reward patenting activity per se, which is not what the system was designed for. Improved prior-art searching could also result in less duplication in research, and/or lead to improved collaboration. As regards solution search, the authors of the paper believe that much better use could be made of the existing literature to find solutions from non-obvious areas of science and technology. The so-called cross industry innovation could be joined by biomimetics, the
inspiration of solutions from nature. Crowdsourcing the concept shorthand could produce a system ‘by the people, for
the people’, to quote Abraham Lincoln out of context. A Citizen Science and Technology initiative that developed a working search engine could generate revenue for academia. Any monies accruing could be invested in research for the common good, such as the development of climate change mitigation technologies, or the discovery of new antibiotics
A search for soft X-ray emission associated with prominent high-velocity-cloud complexes
We correlate the ROSAT 1/4 keV all-sky survey with the Leiden/Dwingeloo HI
survey, looking for soft X-ray signatures of prominent high-velocity-cloud
(HVC) complexes. We study the transfer of 1/4 keV photons through the
interstellar medium in order to distinguish variations in the soft X-ray
background (SXRB) intensity caused by photoelectric absorption effects from
those due to excess X-ray emission. The X-ray data are modelled as a
combination of emission from the Local Hot Bubble (LHB) and emission from a
distant plasma in the galactic halo and extragalactic sources. The X-ray
radiation intensity of the galactic halo and extragalactic X-ray background is
modulated by the photoelectric absorption of the intervening galactic
interstellar matter. We show that large- and small-scale intensity variations
of the 1/4 keV SXRB are caused by photoelectric absorption which is
predominantly traced by the total N(HI) distribution. The extensive coverage of
the two surveys supports evidence for a hot, X-ray emitting corona. We show
that this leads to a good representation of the SXRB observations. For four
large areas on the sky, we search for regions where the modelled and observed
X-ray emission differ. We find that there is excess X-ray emission towards
regions near HVC complexes C, D, and GCN. We suggest that the excess X-ray
emission is positionally correlated with the high-velocity clouds. Some lines
of sight towards HVCs also pass through significant amounts of
intermediate-velocity gas, so we cannot constrain the possible role played by
IVC gas in these directions of HVC and IVC overlap, in determining the X-ray
excesses.Comment: 16 pages, 8 figures, accepted for publication in Astronomy &
Astrophysics main journa
The Leiden/Argentine/Bonn (LAB) Survey of Galactic HI: Final data release of the combined LDS and IAR surveys with improved stray-radiation corrections
We present the final data release of observations of lambda 21-cm emission
from Galactic neutral hydrogen over the entire sky, merging the
Leiden/Dwingeloo Survey (LDS: Hartmann & Burton, 1997) of the sky north of
delta = -30 deg with the Instituto Argentino de Radioastronomia Survey (IAR:
Arnal et al., 2000, and Bajaja et al., 2005) of the sky south of delta = -25
deg. The angular resolution of the combined material is HPBW ~ 0.6 deg. The LSR
velocity coverage spans the interval -450 km/s to +400 km/s, at a resolution of
1.3 km/s. The data were corrected for stray radiation at the Institute for
Radioastronomy of the University of Bonn, refining the original correction
applied to the LDS. The rms brightness-temperature noise of the merged database
is 0.07 - 0.09 K. Residual errors in the profile wings due to defects in the
correction for stray radiation are for most of the data below a level of 20 -
40 mK. It would be necessary to construct a telescope with a main beam
efficiency of eta_{MB} > 99% to achieve the same accuracy. The merged and
refined material entering the LAB Survey of Galactic HI is intended to be a
general resource useful to a wide range of studies of the physical and
structural characteristices of the Galactic interstellar environment. The LAB
Survey is the most sensitive Milky Way HI survey to date, with the most
extensive coverage both spatially and kinematically.Comment: 9 pages, 5 figures, accepted for publication by Astronomy &
Astrophysic
The Leiden/Argentine/Bonn (LAB) survey of Galactic HI : Final data release of the combined IDS and IAR surveys with improved stray-radiation corrections
We present the final data release of observations of λ21-cm emission from Galactic neutral hydrogen over the entire sky, merging the Leiden/Dwingeloo Survey (LDS: Hartmann & Burton 1997, Atlas of Galactic Neutral Hydrogen) of the sky north of δ = -30° with the Instituto Argentino de Radioastronomía Survey (IAR: Arnal et al. 2000, A&AS, 142, 35; and Bajaja et al. 2005, A&A, 440, 767) of the sky south of δ= -25°. The angular resolution of the combined material is HPBW ∼ 0°.6. The LSR velocity coverage spans the interval -450 km s-1 to +400 km s-1, at a resolution of 1.3 km s-1. The data were corrected for stray radiation at the Institute for Radioastronomy of the University of Bonn, refining the original correction applied to the LDS. The rms brightness-temperature noise of the merged database is 0.07-0.09 K. Residual errors in the profile wings due to defects in the correction for stray radiation are for most of the data below a level of 20-40 mK. It would be necessary to construct a telescope with a main beam efficiency of ηMB ≳ 99% to achieve the same accuracy. The merged and refined material entering the LAB Survey of Galactic H I is intended to be a general resource useful to a wide range of studies of the physical and structural characteristices of the Galactic interstellar environment. The LAB Survey is the most sensitive Milky Way H I survey to date, with the most extensive coverage both spatially and kinematically.Instituto Argentino de RadioastronomíaFacultad de Ciencias Astronómicas y Geofísica
- …