7,519 research outputs found

    CHORUS Deliverable 4.5: Report of the 3rd CHORUS Conference

    Get PDF
    The third and last CHORUS conference on Multimedia Search Engines took place from the 26th to the 27th of May 2009 in Brussels, Belgium. About 100 participants from 15 European countries, the US, Japan and Australia learned about the latest developments in the domain. An exhibition of 13 stands presented 16 research projects currently ongoing around the world

    Web crawler research methodology

    Get PDF
    In economic and social sciences it is crucial to test theoretical models against reliable and big enough databases. The general research challenge is to build up a well-structured database that suits well to the given research question and that is cost efficient at the same time. In this paper we focus on crawler programs that proved to be an effective tool of data base building in very different problem settings. First we explain how crawler programs work and illustrate a complex research process mapping business relationships using social media information sources. In this case we illustrate how search robots can be used to collect data for mapping complex network relationship to characterize business relationships in a well defined environment. After that extend the case and present a framework of three structurally different research models where crawler programs can be applied successfully: exploration, classification and time series analysis. In the case of exploration we present findings about the Hungarian web agency industry when no previous statistical data was available about their operations. For classification we show how the top visited Hungarian web domains can be divided into predefined categories of e-business models. In the third research we used a crawler to gather the values of concrete pre-defined records containing ticket prices of low cost airlines from one single site. Based on the experiences we highlight some conceptual conclusions and opportunities of crawler based research in e-business. --e-business research,web search,web crawler,Hungarian web,social network analyis

    Personalized Web Search Techniques - A Review

    Get PDF
    Searching is one of the commonly used task on the Internet. Search engines are the basic tool of the internet, from which related information can be collected according to the specified query or keyword given by the user, and are extremely popular for recurrently used sites. With the remarkable development of the World Wide Web (WWW), the information search has grown to be a major business segment of a global, competitive and money-making market. A perfect search engine is the one which should travel through all the web pages inthe WWW and should list the related information based on the given user keyword. In spite of the recent developments on web search technologies, there are still many conditions in which search engine users obtains the non-relevant search results from the search engines. A personalized Web search has various levels of efficiency for different users, queries, and search contexts. Even though personalized search has been a major research area for many years and many personalization approaches have been examined, it is still uncertain whether personalization is always significant on different queries for diverse users and under different search contexts. This paper focusses on the survey of many efficient personalized Web search approaches which were proposed by many authors

    The Edgar Internet Project: Web Application Development Considerations

    Get PDF
    This paper describes the implementation of World-Wide Web (WWW) access of the SEC EDGAR (Electronic Data Gathering, Archiving and Retrieval) data base. EDGAR is a large, heterogeneous financial data archive that has been available to Internet users since January 1994. It is composed of all forms filed electronically to the Securities and Exchange Commission (SEC) by domestic publicly traded corporations and mutualfunds. We describe WWW design decisions and problems encountered in implementing a public access system to a large database. Our current applications include: an object-oriented mutual fund equity holdings database, a structured full text index search oncorporate profiles, and real-time graphical visualization of stock price and mutual fund position changes

    Facebook in the Australian News: a corpus linguistic approach

    Get PDF
    This thesis analyses the reporting about Facebook in the Australian newsprint media over time, from 2004 to 2013. Based on linguistic analysis of news values, it investigates how traditional news organisations have presented Facebook as ‘newsworthy’. It makes use of a 104,514 word specialised corpus built specifically for the investigation called the ‘Facebook News Corpus’ (FNC), which consists of Australian news texts that appeared around three main events in the company’s history: 1) the launch of Facebook in Australia on 4 February 2004; 2) the listing of Facebook Inc. on Nasdaq on 18 May 2012; and 3) the introduction of Graph Search on 15 January 2013. The FNC is used to examine how news values are construed around a central topic, representing the first attempt to use corpus linguistics to evaluate news about Facebook. The thesis applies an iterative sequence of corpus linguistic techniques, drawing on quantitative and qualitative methods and analytical frameworks, especially Bednarek and Caple’s (2014) discursive news values analysis (DNVA). The study identifies important news values, clusters of co-occurring news values, and how they are constructed through language. It also provides empirical evidence for shifts in news discourse about Facebook over the three time periods that are investigated. Given the rise of Facebook as a primary news source for its more than two billion users, this information will be useful for future research on the role of social networking sites and their relationship with traditional news organisations
    • …
    corecore