348,595 research outputs found
Recommended from our members
Application of Natural Language Processing and Evidential Analysis to Web-Based Intelligence Information Acquisition
The quality of decisions made in business and government relates directly to the quality of the information used to formulate the decision. This information may be retrieved from an organization's knowledge base (Intranet) or from the World Wide Web. Intelligence services Intranet held information can be efficiently manipulated by technologies based upon either semantics such as ontologies, or statistics such as meaning-based computing. These technologies require complex processing of large amount of textual information. However, they cannot currently be effectively applied to Web-based search due to various obstacles, such as lack of semantic tagging. A new approach proposed in this paper supports Web-based search for intelligence information utilizing evidence-based natural language processing (NLP). This approach combines traditional NLP methods for filtering of Web-search results, Grounded Theory to test the completeness of the evidence, and Evidential Analysis to test the quality of gathered information. The enriched information derived from the Web-search will be transferred to the intelligence services knowledge base for handling by an effective Intranet search system thus increasing substantially the information for intelligence analysis. The paper will show that the quality of retrieved information is significantly enhanced by the discovery of previously unknown facts derived from known facts
Workshop on web information seeking and interaction
The World Wide Web has provided access to a diverse range of information sources and systems. People engaging with this rich network of information may need to interact with different technologies, interfaces, and information providers in the course of a single search task. These systems may offer different interaction affordances and require users to adapt their informationseeking strategies. Not only is this challenging for users, but it also presents challenges for the designers of interactive systems, who need to make their own system useful and usable to broad user groups. The popularity of Web browsing and Web search engines has given rise to distinct forms of information-seeking behaviour, and new interaction styles, but we do not yet fully understand these or their implications for the development of new systems
Applying digital content management to support localisation
The retrieval and presentation of digital content such as that on the World Wide Web (WWW) is a substantial area of research. While recent years have seen huge expansion in the size of web-based archives that can be searched efficiently by commercial search engines, the presentation of potentially relevant content is still limited to ranked document lists represented by simple text snippets or image keyframe surrogates. There is expanding interest in techniques to personalise the presentation of content to improve the richness and effectiveness of the user experience. One of the most significant challenges to achieving this is the increasingly multilingual nature of this data, and the need to provide suitably localised responses to users based on this content. The Digital Content Management (DCM) track of the Centre for Next Generation Localisation (CNGL) is seeking to develop technologies to support advanced personalised access and presentation of information by combining elements from the existing research areas of Adaptive Hypermedia and Information Retrieval. The combination of these technologies is intended to produce significant improvements in the way users access information. We review key features of these technologies and introduce early ideas for how these technologies can support localisation and localised content before concluding with some impressions of future directions in DCM
Семантический способ поиска информационных аномалий через web
Semantic search for information anomalies technique is described. It uses semantic web technologies, recommended by World Wide Web Consortium (W3C) — RDF and OWL languages for semantic data description, SPARQL language for making search queries in semantic data. Prototype of system is implemented; it contains semantic data and queries for searching class of information anomalies. Current search systems are reviewed.В статье описывается подход к семантическому поиску информационных аномалий. Используются технологии семантического web, предложенные World Wide Web Consortium (W3C), а именно языки для описания семантических данных RDF и OWL, а также язык формальных поисковых запросов к семантическим данным SPARQL. Реализован прототип системы, где в качестве примера описаны семантические данные и запросы для поиска класса информационных аномалий. Приведен обзор существующих поисковых систем
Indexing and retrieval in digital libraries : developing taxonomies for a repository of decision technologies
DecisionNet is an online Internet-based repository of decision technologies. It links remote users with these technologies and provides a directory service to enable search and selection of suitable technologies. The ability to retrieve relevant objects through search mechanisms is basic to any repository's success and usability and depends on effective classification of the decision technologies. This thesis develops classification methods to enable indexing of the DecisionNet repository. Existing taxonomies for software and other online repositories are examined. Criteria and principles for a good taxonomy are established and systematically applied to develop DecisionNet taxonomies. A database design is developed to store the taxonomies and to classify the technologies in the repository. User interface issues for navigation of a hierarchical classification system are discussed. A user interface for remote World Wide Web users is developed. This user interface is designed for browsing the taxonomy structure and creating search parameters online. Recommendations for the implementation of a repository search mechanism are given.http://archive.org/details/indexingndretrie1094532199NAU.S. Navy (U.S.N.) authorApproved for public release; distribution is unlimited
Personalized Web Search Techniques - A Review
Searching is one of the commonly used task on the Internet. Search engines are the basic tool of the internet, from which related information can be collected according to the specified query or keyword given by the user, and are extremely popular for recurrently used sites. With the remarkable development of the World Wide Web (WWW), the information search has grown to be a major business segment of a global, competitive and money-making market. A perfect search engine is the one which should travel through all the web pages inthe WWW and should list the related information based on the given user keyword. In spite of the recent developments on web search technologies, there are still many conditions in which search engine users obtains the non-relevant search results from the search engines. A personalized Web search has various levels of efficiency for different users, queries, and search contexts. Even though personalized search has been a major research area for many years and many personalization approaches have been examined, it is still uncertain whether personalization is always significant on different queries for diverse users and under different search contexts. This paper focusses on the survey of many efficient personalized Web search approaches which were proposed by many authors
CS 875: Semantic Web
World Wide Web (Web 1.0, or the Web, as we now know it) centers on documents and semistructured data in html, rss, and xml. The next generation Web, also called Web 2.0 and Web 3.0, has already started to emerge. Web 2.0 is about user-generated content, user participation such as through tagging, and social networking. Web 3.0, also called Semantic Web, is about labeling content such that machines can process it more intelligently and humans can exploit it more effectively. These labels or metadata add semantics (meaning) to data, and their formal representation enables powerful reasoning that leads not only to better (semantic) search but also to analysis, discovery, and decision making. Semantic Web is already a rapidly emerging field, with standards, technologies, products, and applications-as well as to excellent job prospects (for MS students) and research opportunities (for PhD students)
The Present and Future of Internet Search
Search engines were crucial in the development of the World Wide Web. Web-based information retrieval progressed from simple word matching to sophisticated algorithms for maximizing the relevance of search results. Statistical and graph-based approaches for indexing and ranking pages, natural language processing techniques for improving query results, and intelligent agents for personalizing the search process all show great promise for enhanced performance. The evolution in search technology was accompanied by growing economic pressures on search engine companies. Unable to sustain long-term viability from advertising revenues, many of the original search engines diversified into portals that farm out their search and directory operations. Vertical portals that serve focused user communities also outsource their search services, and even directory providers began to integrate search engine technologies from outside vendors. This article brings order to the chaos resulting from the variety of search tools being offered under various marketing guises. While growing reliance on a small set of search providers is leading to less diversity among search services, users can expect individualized searching experiences that factor in personal information. The convergence of technology and business models also results in more narrowly defined search spaces, which will lessen the quantity of search results while improving their quality
Searching the WWW with XML
Searching and retrieving the right information from the World-Wide Web (WWW) has always been considered of foremost importance and of considerable A.I. intensivity. Internet search technologies have been evolving over the years and will continue to do so as the WWW will continue to expand in size and increase in popularity. In a desperate attempt to restore order to the WWW after the chaos that has developed due to its heterogeneous, unstructured and uncensored nature, the eXtended Markup Language (XML) is being heralded as the successor to HTML. In this paper we investigate the evolution of Internet search technologies and present a possible and viable solution in a functional system we developed and which makes use of XML at its very core. We discuss the design issues involved as well as practical issues such as tendencies and tactics employed by some of the major players in this well-sought area
- …