9 research outputs found
On quantum statistics in data analysis
Originally, quantum probability theory was developed to analyze statistical
phenomena in quantum systems, where classical probability theory does not
apply, because the lattice of measurable sets is not necessarily distributive.
On the other hand, it is well known that the lattices of concepts, that arise
in data analysis, are in general also non-distributive, albeit for completely
different reasons. In his recent book, van Rijsbergen argues that many of the
logical tools developed for quantum systems are also suitable for applications
in information retrieval. I explore the mathematical support for this idea on
an abstract vector space model, covering several forms of data analysis
(information retrieval, data mining, collaborative filtering, formal concept
analysis...), and roughly based on an idea from categorical quantum mechanics.
It turns out that quantum (i.e., noncommutative) probability distributions
arise already in this rudimentary mathematical framework. We show that a
Bell-type inequality must be satisfied by the standard similarity measures, if
they are used for preference predictions. The fact that already a very general,
abstract version of the vector space model yields simple counterexamples for
such inequalities seems to be an indicator of a genuine need for quantum
statistics in data analysis.Comment: 7 pages, Quantum Interaction 2008 (Oxford, April 2008) v3: added two
diagrams, changed some wording
An effective approach for personalized web search based on community-cluster analysis
The concept of Personalized Web Search is
commonly used for improving the quality of web search
results by identifying and facilitating different users' search
needs. There are several techniques such as user profiling,
content analysis, hyperlink analysis and biased PageRank
algorithm that are used to achieve web personalization. User
Profiling is one of the widely used techniques for
personalizing web search at large scale. But it contains
several technical and ethical issues such as privacy violations,
inefficient use of computing resources as well. Collaborative
web search is also a kind of a relatively "new concept which
defines the way of optimizing/personalizing search results by
using details of group of people and contributing the
knowledge of all of them about web search. This paper
presents the details of an alternative approach for
personalizing web results by using user profiling technique
with community cluster analysis of collaborative web search
by adapting concept of reusability 'among web results
Networks, complexity and internet regulation: scale-free law
No description supplie
Networks, complexity and internet regulation scale-free law
This book, then, starts with a general statement: that regulators should try, wherever possible, to use the physical methodological tools presently
available in order to draft better legislation. While such an assertion may be applied to
the law in general, this work will concentrate on the much narrower area of Internet
regulation and the science of complex networks The Internet is the subject of this book not only because it is my main area of
research, but also because âwithout over-emphasising the importance of the Internet to
everyday lifeâ one cannot deny that the growth and popularisation of the global
communications network has had a tremendous impact on the way in which we interact
with one another. The Internet is, however, just one of many interactive networks. One
way of looking at the complex and chaotic nature of society is to see it as a collection of
different nodes of interaction. Humans are constantly surrounded by networks: the social
network, the financial network, the transport network, the telecommunications network
and even the network of our own bodies. Understanding how these systems operate and
interact with one another has been the realm of physicists, economists, biologists and
mathematicians. Until recently, the study of networks has been mainly theoretical and
academic, because it is difficult to gather data about large and complex systems that is
sufficiently reliable to support proper empirical application. In recent years, though, the
Internet has given researchers the opportunity to study and test the mathematical
descriptions of these vast complex systems. The growth rate and structure of cyberspace
has allowed researchers to map and test several previously unproven theories about how
links and hubs within networks interact with one another. The Web now provides the
means with which to test the organisational structures, architecture and growth of
networks, and even permits some limited prediction about their behaviour, strengths and
vulnerabilities. The main objective of this book is first and foremost to serve as an introduction to the
wider legal audience to some of the theories of complexity and networks. The second
objective is more ambitious. By looking at the application of complexity theory and network science in various areas of Internet regulation, it is hoped that there will be
enough evidence to postulate a theory of Internet regulation based on network science.
To achieve these two goals, Chapter 2 will look in detail at the science of complex
networks to set the stage for the legal and regulatory arguments to follow. With the
increase in reliability of the descriptive (and sometimes predictive) nature of network
science, a logical next step for legal scholars is to look at the legal implications of the
characteristics of networks. Chapter 3 highlights the efforts of academics and
practitioners who have started to find potential uses for network science tools. Chapter 4
takes this idea further, and explores how network theory can shape Internet regulation.
The following chapters will analyse the potential for application of the tools described
in the previous chapters, applying complexity theory to specific areas of study related to
Internet Law. Chapter 5 deals with the subject of copyright in the digital world. Chapter
6 explores the issue of peer-production and user-generated content using network
science as an analytical framework. Chapter 7 finishes the evidence section of the work
by studying the impact of network architecture in the field of cybercrime, and asks
whether the existing architecture hinders or assists efforts to tackle those problems.
It is clear that these are very disparate areas of study. It is not the intention of this
book to be overreaching in its scope, although I am mindful that it covers a lot of ground
and attempts to study and describe some disciplines that fall outside of my intellectual
comfort zone. While the focus of the work is the Internet, its applications may extend
beyond mere electronic bits. Without trying to be over-ambitious, it is my strong belief
that legal scholarship has been neglectful in that it has been slow to respond to the
wealth of research into complexity. That is not to say that there has been no legal
research on the topic, but it would seem that lawyers, legislators and policy-makers are
reluctant to consider technical solutions to legal problems. It is hoped then that this work
will serve as a stepping stone that will lead to new interest in some of the theories that I
describe
Deriving and applying facet views of the Dewey Decimal Classification Scheme to enhance subject searching in library OPACs
Classification is a fundamental tool in the organisation of any library collection for effective information retrieval. Several classifications exist, yet the pioneering Dewey Decimal Classification (DDC) still constitutes the most widely used scheme and international de facto standard. Although once used for the dual purpose of physical organisation and subject retrieval in the printed library catalogue, library classification is now relegated to a singular role of shelf location. Numerous studies have highlighted the problem of subject access in library online public access catalogues (OPACs). The library OPAC has changed relatively little since its inception, designed to find what is already known, not discover and explore. This research aims to enhance OPAC subject searching by deriving facets of the DDC and populating these with a library collection for display at a View-based searching OPAC interface. A novel method is devised that enables the automatic deconstruction of complex DDC notations into their component facets. Identifying facets based upon embedded notational components reveals alternative, multidimensional subject arrangements of a library collection and resolves the problem of disciplinary scatter. The extent to which the derived facets enhance users' subject searching perceptions and activities at the OPAC interface is evaluated in a small-scale usability study. The results demonstrate the successful derivation of four fundamental facets (Reference Type, Person Type, Time and Geographic Place). Such facet derivation and deconstruction of Dewey notations is recognised as a complex process, owing to the lack of a uniform notation, notational re-use and the need for distinct facet indicators to delineate facet boundaries. The results of the preliminary usability study indicate that users are receptive to facet-based searching and that the View-based searching system performs equally as well as a current form fill-in interface and, in some cases, provides enhanced benefits. It is concluded that further exploration of facet-based searching is clearly warranted and suggestions for future research are made.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
Recommended from our members
Information in nineteenth century England : exploring contemporary socio-cultural perceptions and understandings
Historical interest in the information society and the tools and techniques of information processing and management has been the subject of much recent scholarship. In contrast, this research examines understandings of information from a social and cultural perspective and offers new insights into the chronology of `modern' information. The focus is on nineteenth century England and how contemporary society thought of and understood information. As a consequence of this, this work also investigates the role and manifestation of information according to these perceptions. The research explores perceptions from a variety of nineteenth century viewpoints, including etiquette and behavioural literature, Victorian periodicals, events such as the Channel Tunnel panics of the 1880s and discrete individual case studies. The thesis concludes that the nineteenth century formed the crux point between premodem understandings of information, and the start of what we would now term a period of modernity so far as information is concerned. The foundation of the research is nineteenth century published and archival material, contemporary newspapers, journals and private papers