96 research outputs found
Sampling properties of directed networks
For many real-world networks only a small "sampled" version of the original
network may be investigated; those results are then used to draw conclusions
about the actual system. Variants of breadth-first search (BFS) sampling, which
are based on epidemic processes, are widely used. Although it is well
established that BFS sampling fails, in most cases, to capture the
IN-component(s) of directed networks, a description of the effects of BFS
sampling on other topological properties are all but absent from the
literature. To systematically study the effects of sampling biases on directed
networks, we compare BFS sampling to random sampling on complete large-scale
directed networks. We present new results and a thorough analysis of the
topological properties of seven different complete directed networks (prior to
sampling), including three versions of Wikipedia, three different sources of
sampled World Wide Web data, and an Internet-based social network. We detail
the differences that sampling method and coverage can make to the structural
properties of sampled versions of these seven networks. Most notably, we find
that sampling method and coverage affect both the bow-tie structure, as well as
the number and structure of strongly connected components in sampled networks.
In addition, at low sampling coverage (i.e. less than 40%), the values of
average degree, variance of out-degree, degree auto-correlation, and link
reciprocity are overestimated by 30% or more in BFS-sampled networks, and only
attain values within 10% of the corresponding values in the complete networks
when sampling coverage is in excess of 65%. These results may cause us to
rethink what we know about the structure, function, and evolution of real-world
directed networks.Comment: 21 pages, 11 figure
Index ordering by query-independent measures
Conventional approaches to information retrieval search through all applicable entries in an inverted file for a particular collection in order to find those documents with the highest scores. For particularly large collections this may be extremely time consuming.
A solution to this problem is to only search a limited amount of the collection at query-time, in order to speed up the retrieval process. In doing this we can also limit the loss in retrieval efficacy (in terms of accuracy of results). The way we achieve this is to firstly identify the most “important” documents within the collection, and sort documents within inverted file lists in order of this “importance”. In this way we limit the amount of information to be searched at query time by eliminating documents of lesser importance, which not only makes the search more efficient, but also limits loss in retrieval accuracy. Our experiments, carried out on the TREC Terabyte collection, report significant savings, in terms of number of postings examined, without significant loss of effectiveness when based on several measures of importance used in isolation, and in combination. Our results point to several ways in which the computation cost of searching large collections of documents can be significantly reduced
Homophily in the Digital World: A LiveJournal Case Study
Special issue on Social Computing in Blogosphere (IC)</p
- …