80 research outputs found

    Crowd-Sourcing Fuzzy and Faceted Classification for Concept Search

    Full text link
    Searching for concepts in science and technology is often a difficult task. To facilitate concept search, different types of human-generated metadata have been created to define the content of scientific and technical disclosures. Classification schemes such as the International Patent Classification (IPC) and MEDLINE's MeSH are structured and controlled, but require trained experts and central management to restrict ambiguity (Mork, 2013). While unstructured tags of folksonomies can be processed to produce a degree of structure (Kalendar, 2010; Karampinas, 2012; Sarasua, 2012; Bragg, 2013) the freedom enjoyed by the crowd typically results in less precision (Stock 2007). Existing classification schemes suffer from inflexibility and ambiguity. Since humans understand language, inference, implication, abstraction and hence concepts better than computers, we propose to harness the collective wisdom of the crowd. To do so, we propose a novel classification scheme that is sufficiently intuitive for the crowd to use, yet powerful enough to facilitate search by analogy, and flexible enough to deal with ambiguity. The system will enhance existing classification information. Linking up with the semantic web and computer intelligence, a Citizen Science effort (Good, 2013) would support innovation by improving the quality of granted patents, reducing duplicitous research, and stimulating problem-oriented solution design. A prototype of our design is in preparation. A crowd-sourced fuzzy and faceted classification scheme will allow for better concept search and improved access to prior art in science and technology

    Durham Zoo: powering a search-&-innovation engine with collective intelligence

    Get PDF
    Purpose – Durham Zoo (hereinafter – DZ) is a project to design and operate a concept search engine for science and technology. In DZ, a concept includes a solution to a problem in a particular context. Design – Concept searching is rendered complex by the fuzzy nature of a concept, the many possible implementations of the same concept, and the many more ways that the many implementations can be expressed in natural language. An additional complexity is the diversity of languages and formats, in which the concepts can be disclosed. Humans understand language, inference, implication and abstraction and, hence, concepts much better than computers, that in turn are much better at storing and processing vast amounts of data. We are 7 billion on the planet and we have the Internet as the backbone for Collective Intelligence. So, our concept search engine uses humans to store concepts via a shorthand that can be stored, processed and searched by computers: so, humans IN and computers OUT. The shorthand is classification: metadata in a structure that can define the content of a disclosure. The classification is designed to be powerful in terms of defining and searching concepts, whilst suited to a crowdsourcing effort. It is simple and intuitive to use. Most importantly, it is adapted to restrict ambiguity, which is the poison of classification, without imposing a restrictive centralised management. In the classification scheme, each entity is shown together in a graphical representation with related entities. The entities are arranged on a sliding scale of similarity. This sliding scale is effectively fuzzy classification. Findings – The authors of the paper have been developing a first classification scheme for the technology of traffic cones, this in preparation for a trial of a working system. The process has enabled the authors to further explore the practicalities of concept classification. The CmapTools knowledge modelling kit to develop the graphical representations has been used. Practical implications – Concept searching is seen as having two categories: prior art searching, which is searching for what already exists, and solution searching: a search for a novel solution to an existing problem. Prior art searching is not as efficient a process, as all encompassing in scope, or as accurate in result, as it could and probably should be. The prior art includes library collections, journals, conference proceedings and everything else that has been written, drawn, spoken or made public in any way. Much technical information is only published in patents. There is a good reason to improve prior art searching: research, industry, and indeed humanity faces the spectre of patent thickets: an impenetrable legal space that effectively hinders innovation rather than promotes it. Improved prior-art searching would help with the gardening and result in fewer and higher-quality patents. Poor-quality patents can reward patenting activity per se, which is not what the system was designed for. Improved prior-art searching could also result in less duplication in research, and/or lead to improved collaboration. As regards solution search, the authors of the paper believe that much better use could be made of the existing literature to find solutions from non-obvious areas of science and technology. The so-called cross industry innovation could be joined by biomimetics, the inspiration of solutions from nature. Crowdsourcing the concept shorthand could produce a system ‘by the people, for the people’, to quote Abraham Lincoln out of context. A Citizen Science and Technology initiative that developed a working search engine could generate revenue for academia. Any monies accruing could be invested in research for the common good, such as the development of climate change mitigation technologies, or the discovery of new antibiotics

    A search for soft X-ray emission associated with prominent high-velocity-cloud complexes

    Get PDF
    We correlate the ROSAT 1/4 keV all-sky survey with the Leiden/Dwingeloo HI survey, looking for soft X-ray signatures of prominent high-velocity-cloud (HVC) complexes. We study the transfer of 1/4 keV photons through the interstellar medium in order to distinguish variations in the soft X-ray background (SXRB) intensity caused by photoelectric absorption effects from those due to excess X-ray emission. The X-ray data are modelled as a combination of emission from the Local Hot Bubble (LHB) and emission from a distant plasma in the galactic halo and extragalactic sources. The X-ray radiation intensity of the galactic halo and extragalactic X-ray background is modulated by the photoelectric absorption of the intervening galactic interstellar matter. We show that large- and small-scale intensity variations of the 1/4 keV SXRB are caused by photoelectric absorption which is predominantly traced by the total N(HI) distribution. The extensive coverage of the two surveys supports evidence for a hot, X-ray emitting corona. We show that this leads to a good representation of the SXRB observations. For four large areas on the sky, we search for regions where the modelled and observed X-ray emission differ. We find that there is excess X-ray emission towards regions near HVC complexes C, D, and GCN. We suggest that the excess X-ray emission is positionally correlated with the high-velocity clouds. Some lines of sight towards HVCs also pass through significant amounts of intermediate-velocity gas, so we cannot constrain the possible role played by IVC gas in these directions of HVC and IVC overlap, in determining the X-ray excesses.Comment: 16 pages, 8 figures, accepted for publication in Astronomy & Astrophysics main journa

    The Leiden/Argentine/Bonn (LAB) Survey of Galactic HI: Final data release of the combined LDS and IAR surveys with improved stray-radiation corrections

    Get PDF
    We present the final data release of observations of lambda 21-cm emission from Galactic neutral hydrogen over the entire sky, merging the Leiden/Dwingeloo Survey (LDS: Hartmann & Burton, 1997) of the sky north of delta = -30 deg with the Instituto Argentino de Radioastronomia Survey (IAR: Arnal et al., 2000, and Bajaja et al., 2005) of the sky south of delta = -25 deg. The angular resolution of the combined material is HPBW ~ 0.6 deg. The LSR velocity coverage spans the interval -450 km/s to +400 km/s, at a resolution of 1.3 km/s. The data were corrected for stray radiation at the Institute for Radioastronomy of the University of Bonn, refining the original correction applied to the LDS. The rms brightness-temperature noise of the merged database is 0.07 - 0.09 K. Residual errors in the profile wings due to defects in the correction for stray radiation are for most of the data below a level of 20 - 40 mK. It would be necessary to construct a telescope with a main beam efficiency of eta_{MB} > 99% to achieve the same accuracy. The merged and refined material entering the LAB Survey of Galactic HI is intended to be a general resource useful to a wide range of studies of the physical and structural characteristices of the Galactic interstellar environment. The LAB Survey is the most sensitive Milky Way HI survey to date, with the most extensive coverage both spatially and kinematically.Comment: 9 pages, 5 figures, accepted for publication by Astronomy & Astrophysic

    The Leiden/Argentine/Bonn (LAB) survey of Galactic HI : Final data release of the combined IDS and IAR surveys with improved stray-radiation corrections

    Get PDF
    We present the final data release of observations of λ21-cm emission from Galactic neutral hydrogen over the entire sky, merging the Leiden/Dwingeloo Survey (LDS: Hartmann & Burton 1997, Atlas of Galactic Neutral Hydrogen) of the sky north of δ = -30° with the Instituto Argentino de Radioastronomía Survey (IAR: Arnal et al. 2000, A&AS, 142, 35; and Bajaja et al. 2005, A&A, 440, 767) of the sky south of δ= -25°. The angular resolution of the combined material is HPBW ∼ 0°.6. The LSR velocity coverage spans the interval -450 km s-1 to +400 km s-1, at a resolution of 1.3 km s-1. The data were corrected for stray radiation at the Institute for Radioastronomy of the University of Bonn, refining the original correction applied to the LDS. The rms brightness-temperature noise of the merged database is 0.07-0.09 K. Residual errors in the profile wings due to defects in the correction for stray radiation are for most of the data below a level of 20-40 mK. It would be necessary to construct a telescope with a main beam efficiency of ηMB ≳ 99% to achieve the same accuracy. The merged and refined material entering the LAB Survey of Galactic H I is intended to be a general resource useful to a wide range of studies of the physical and structural characteristices of the Galactic interstellar environment. The LAB Survey is the most sensitive Milky Way H I survey to date, with the most extensive coverage both spatially and kinematically.Instituto Argentino de RadioastronomíaFacultad de Ciencias Astronómicas y Geofísica
    corecore