925 research outputs found

    Expertise-based peer selection in Peer-to-Peer networks

    Get PDF
    Peer-to-Peer systems have proven to be an effective way of sharing data. Modern protocols are able to efficiently route a message to a given peer. However, determining the destination peer in the first place is not always trivial. We propose a a message to a given peer. However, determining the destination peer in the first place is not always trivial. We propose a model in which peers advertise their expertise in the Peer-to-Peer network. The knowledge about the expertise of other peers forms a semantic topology. Based on the semantic similarity between the subject of a query and the expertise of other peers, a peer can select appropriate peers to forward queries to, instead of broadcasting the query or sending it to a random set of peers. To calculate our semantic similarity measure, we make the simplifying assumption that the peers share the same ontology. We evaluate the model in a bibliographic scenario, where peers share bibliographic descriptions of publications among each other. In simulation experiments complemented with a real-world field experiment, we show how expertise-based peer selection improves the performance of a Peer-to-Peer system with respect to precision, recall and the number of messages

    Semantic Flooding: Semantic Search across Distributed Lightweight Ontologies

    Get PDF
    Lightweight ontologies are trees where links between nodes codify the fact that a node lower in the hierarchy describes a topic (and contains documents about this topic) which is more specific than the topic of the node one level above. In turn, multiple lightweight ontologies can be connected by semantic links which represent mappings among them and which can be computed, e.g., by ontology matching. In this paper we describe how these two types of links can be used to define a semantic overlay network which can cover any number of peers and which can be flooded to perform a semantic search on documents, i.e., to perform semantic flooding. We have evaluated our approach by simulating a network of 10,000 peers containing classifications which are fragments of the DMoz web directory. The results are promising and show that, in our approach, only a relatively small number of peers needs to be queried in order to achieve high accuracy

    ZBroker: A query routing broker for Z39.50 databases

    Get PDF
    A query routing broker is a software agent that determines from a large set of accessing information sources the ones most relevant to a user's information need. As the number of information sources on the Internet increases dramatically, future users will have to rely on query routing brokers to decide a small number of information sources to query without incurring too much query processing overheads. In this paper, we describe a query routing broker known as ZBroker developed for bibliographic database servers that support the Z39.50 protocol. ZBroker samples the content of each bibliographic database by using training queries and their results, and summarizes the bibliographic database content into a knowledge base. We present the design and implementation of ZBroker and describe its Web-based user interface.Comment: 10 pages, 8 figure

    Ontology engineering and routing in distributed knowledge management applications

    Get PDF

    Implementation of Paper Genealogy in Subgraph Mining

    Get PDF
    Information networks contains many data base in the different search of area , Whenever a new researcher goes to search a topic , there are lots of papers , In those papers some are relevant to user define topic and some are unfamiliar to that topic.For making literature survey researcher needs to collect all information regarding domain which are relevant to that particular topic but there are many citations are available which contains huge amount of data where number of papersis presented by authors.It is very difficult to study all published papers, after analysing this problem an idea is created to solve the problem of search of all research papers with their citation. This paper is design to solve these entire problems, how to find out relative papers with respected query. This paper will be centred on creation of genealogy of all those published papers which will find out the all relevant papers according to user entered keyword it is startingworking of process, after that extraction part will be come in which discrimination of survey paper and implementation of paper will be extracting according to seminal papers it will create genealogy of those paper, by association and interlinking among all matching documents on the basis of references of each paper. The created Genealogy willhelpful for user to get a quick look of their searched topic at which papers are relevant to given query of research, So that all the seminal papers will be shown to user and usercan focus on only those documents . By this proposed work user neither looks on unwanted documents nor expend the time for searching the particular topic, which may increases scalability and efficiency of searching keywords

    Improving the efficiency of spam filtering through cache architecture

    Get PDF
    Blacklists (BLs), also called Domain Name Systembased Blackhole List (DNSBLs) are the databases of known internet addresses used by the spammers to send out the spam mails. Mail servers use these lists to filter out the e-mails coming from different spam sources. In contrary, Whitelists (WLs) are the explicit list of senders from whom e-mail can be accepted or delivered. Mail Transport Agent (MTA) is usually configured to reject, challenge or flag the messages which have been sent from the sources listed on one or more DNSBLs and to allow the messages from the sources listed on the WLs. In this paper, we are demonstrating how the bandwidth (the overall requests and responses that need to go over the network) performance is improved by using local caches for BLs and WLs. The actual sender\u27s IP addresses are extracted from the e-mail log. These are then compared with the list in the local caches to find out if they should be accepted or not, before they are checked against the global DNSBLs by running \u27DNSBL queries\u27 (if required). Around three quarters of the e-mail sources have been observed to be filtered locally through caches with this method. Provision of local control over the lists and lower search (filtering) time are the other related benefits. © 2008 IEEE
    • …
    corecore