2,252 research outputs found
Towards a Model of Understanding Social Search
Search engine researchers typically depict search as the solitary activity of
an individual searcher. In contrast, results from our critical-incident survey
of 150 users on Amazon's Mechanical Turk service suggest that social
interactions play an important role throughout the search process. Our main
contribution is that we have integrated models from previous work in
sensemaking and information seeking behavior to present a canonical social
model of user activities before, during, and after search, suggesting where in
the search process even implicitly shared information may be valuable to
individual searchers.Comment: Presented at 1st Intl Workshop on Collaborative Information Seeking,
2008 (arXiv:0908.0583
An item/user representation for recommender systems based on bloom filters
This paper focuses on the items/users representation
in the domain of recommender systems. These systems compute
similarities between items (and/or users) to recommend new items to users based on their previous preferences. It is often useful to consider the characteristics (a.k.a features or attributes) of the items and/or users. This represents items/users by vectors that can be very large, sparse and space-consuming. In this paper, we propose a new accurate method for representing items/users with low size data structures that relies on two concepts: (1) item/user representation is based on bloom filter vectors, and (2) the usage of these filters to compute bitwise AND similarities and bitwise XNOR similarities. This work is motivated by three ideas: (1) detailed vector representations are large and sparse, (2) comparing more features of items/users may achieve better accuracy for items similarities, and (3) similarities are not only in common existing aspects, but also in common missing aspects.
We have experimented this approach on the publicly available
MovieLens dataset. The results show a good performance in
comparison with existing approaches such as standard vector
representation and Singular Value Decomposition (SVD)
Exploring the Googlesphere: Information Literacy and First-Year College Students
First-year students often lack the information literacy skills necessary to thrive in the 21st century academic environment. The purpose of this pilot study was to explore the impact of a Google literacy intervention on the research skills and attitudes of first-year students enrolled in a first-year seminar course. A secondary purpose of this study is to determine the effectiveness and appropriateness of the Google literacy intervention (“Exploring the Googlesphere”) materials, activities, and assessments. This study used the action research model which provided the researcher with a systematic way to test her assumptions and experiences and to improve the quality of teaching and learning in her own classroom (Kember, 2000; Klipfel & Carroll; 2015). Study participants completed a Google Search pre- and post-test survey, an adapted Critical Incident Questionnaire, and pre- and post-class drawings. These instruments were analyzed using qualitative data analysis. The findings indicated that the intervention was a moderate success, particularly in helping students to develop concrete search skills, conceptual understandings of how Google search operates, and confidence in themselves as researchers and Google as a tool for research. Recommendations for future research and implications for practice are discussed
Why (and How) Networks Should Run Themselves
The proliferation of networked devices, systems, and applications that we
depend on every day makes managing networks more important than ever. The
increasing security, availability, and performance demands of these
applications suggest that these increasingly difficult network management
problems be solved in real time, across a complex web of interacting protocols
and systems. Alas, just as the importance of network management has increased,
the network has grown so complex that it is seemingly unmanageable. In this new
era, network management requires a fundamentally new approach. Instead of
optimizations based on closed-form analysis of individual protocols, network
operators need data-driven, machine-learning-based models of end-to-end and
application performance based on high-level policy goals and a holistic view of
the underlying components. Instead of anomaly detection algorithms that operate
on offline analysis of network traces, operators need classification and
detection algorithms that can make real-time, closed-loop decisions. Networks
should learn to drive themselves. This paper explores this concept, discussing
how we might attain this ambitious goal by more closely coupling measurement
with real-time control and by relying on learning for inference and prediction
about a networked application or system, as opposed to closed-form analysis of
individual protocols
LSST: from Science Drivers to Reference Design and Anticipated Data Products
(Abridged) We describe here the most ambitious survey currently planned in
the optical, the Large Synoptic Survey Telescope (LSST). A vast array of
science will be enabled by a single wide-deep-fast sky survey, and LSST will
have unique survey capability in the faint time domain. The LSST design is
driven by four main science themes: probing dark energy and dark matter, taking
an inventory of the Solar System, exploring the transient optical sky, and
mapping the Milky Way. LSST will be a wide-field ground-based system sited at
Cerro Pach\'{o}n in northern Chile. The telescope will have an 8.4 m (6.5 m
effective) primary mirror, a 9.6 deg field of view, and a 3.2 Gigapixel
camera. The standard observing sequence will consist of pairs of 15-second
exposures in a given field, with two such visits in each pointing in a given
night. With these repeats, the LSST system is capable of imaging about 10,000
square degrees of sky in a single filter in three nights. The typical 5
point-source depth in a single visit in will be (AB). The
project is in the construction phase and will begin regular survey operations
by 2022. The survey area will be contained within 30,000 deg with
, and will be imaged multiple times in six bands, ,
covering the wavelength range 320--1050 nm. About 90\% of the observing time
will be devoted to a deep-wide-fast survey mode which will uniformly observe a
18,000 deg region about 800 times (summed over all six bands) during the
anticipated 10 years of operations, and yield a coadded map to . The
remaining 10\% of the observing time will be allocated to projects such as a
Very Deep and Fast time domain survey. The goal is to make LSST data products,
including a relational database of about 32 trillion observations of 40 billion
objects, available to the public and scientists around the world.Comment: 57 pages, 32 color figures, version with high-resolution figures
available from https://www.lsst.org/overvie
- …