201,336 research outputs found
A new theory of space syntax
Relations between different components of urban structure are often measured in aliteral manner, along streets for example, the usual representation being routesbetween junctions which form the nodes of an equivalent planar graph. A popularvariant on this theme ? space syntax ? treats these routes as streets containing one ormore junctions, with the equivalent graph representation being more abstract, basedon relations between the streets which themselves are treated as nodes. In this paper,we articulate space syntax as a specific case of relations between any two sets, in thiscase, streets and their junctions, from which we derive two related representations.The first or primal problem is traditional space syntax based on relations betweenstreets through their junctions; the second or dual problem is the more usualmorphological representation of relations between junctions through their streets.The unifying framework that we propose suggests we shift our focus from the primalproblem where accessibility or distance is associated with lines or streets, to the dualproblem where accessibility is associated with points or junctions. This traditionalrepresentation of accessibility between points rather than between lines is easier tounderstand and makes more sense visually. Our unifying framework enables us toeasily shift from the primal problem to the dual and back, thus providing a muchricher interpretation of the syntax. We develop an appropriate algebra which providesa clearer approach to connectivity and distance in the equivalent graphrepresentations, and we then demonstrate these variants for the primal and dualproblems in one of the first space syntax street network examples, the French villageof Gassin. An immediate consequence of our analysis is that we show how the directconnectivity of streets (or junctions) to one another is highly correlated with thedistance measures used. This suggests that a simplified form of syntax can beoperationalized through counts of streets and junctions in the original street network
Approximated and User Steerable tSNE for Progressive Visual Analytics
Progressive Visual Analytics aims at improving the interactivity in existing
analytics techniques by means of visualization as well as interaction with
intermediate results. One key method for data analysis is dimensionality
reduction, for example, to produce 2D embeddings that can be visualized and
analyzed efficiently. t-Distributed Stochastic Neighbor Embedding (tSNE) is a
well-suited technique for the visualization of several high-dimensional data.
tSNE can create meaningful intermediate results but suffers from a slow
initialization that constrains its application in Progressive Visual Analytics.
We introduce a controllable tSNE approximation (A-tSNE), which trades off speed
and accuracy, to enable interactive data exploration. We offer real-time
visualization techniques, including a density-based solution and a Magic Lens
to inspect the degree of approximation. With this feedback, the user can decide
on local refinements and steer the approximation level during the analysis. We
demonstrate our technique with several datasets, in a real-world research
scenario and for the real-time analysis of high-dimensional streams to
illustrate its effectiveness for interactive data analysis
Visualising the structure of document search results: A comparison of graph theoretic approaches
This is the post-print of the article - Copyright @ 2010 Sage PublicationsPrevious work has shown that distance-similarity visualisation or ‘spatialisation’ can provide a potentially useful context in which to browse the results of a query search, enabling the user to adopt a simple local foraging or ‘cluster growing’ strategy to navigate through the retrieved document set. However, faithfully mapping feature-space models to visual space can be problematic owing to their inherent high dimensionality and non-linearity. Conventional linear approaches to dimension reduction tend to fail at this kind of task, sacrificing local structural in order to preserve a globally optimal mapping. In this paper the clustering performance of a recently proposed algorithm called isometric feature mapping (Isomap), which deals with non-linearity by transforming dissimilarities into geodesic distances, is compared to that of non-metric multidimensional scaling (MDS). Various graph pruning methods, for geodesic distance estimation, are also compared. Results show that Isomap is significantly better at preserving local structural detail than MDS, suggesting it is better suited to cluster growing and other semantic navigation tasks. Moreover, it is shown that applying a minimum-cost graph pruning criterion can provide a parameter-free alternative to the traditional K-neighbour method, resulting in spatial clustering that is equivalent to or better than that achieved using an optimal-K criterion
Learning Graph Embeddings from WordNet-based Similarity Measures
We present path2vec, a new approach for learning graph embeddings that relies
on structural measures of pairwise node similarities. The model learns
representations for nodes in a dense space that approximate a given
user-defined graph distance measure, such as e.g. the shortest path distance or
distance measures that take information beyond the graph structure into
account. Evaluation of the proposed model on semantic similarity and word sense
disambiguation tasks, using various WordNet-based similarity measures, show
that our approach yields competitive results, outperforming strong graph
embedding baselines. The model is computationally efficient, being orders of
magnitude faster than the direct computation of graph-based distances.Comment: Accepted to StarSem 201
A comparison of two techniques for bibliometric mapping: Multidimensional scaling and VOS
VOS is a new mapping technique that can serve as an alternative to the
well-known technique of multidimensional scaling. We present an extensive
comparison between the use of multidimensional scaling and the use of VOS for
constructing bibliometric maps. In our theoretical analysis, we show the
mathematical relation between the two techniques. In our experimental analysis,
we use the techniques for constructing maps of authors, journals, and keywords.
Two commonly used approaches to bibliometric mapping, both based on
multidimensional scaling, turn out to produce maps that suffer from artifacts.
Maps constructed using VOS turn out not to have this problem. We conclude that
in general maps constructed using VOS provide a more satisfactory
representation of a data set than maps constructed using well-known
multidimensional scaling approaches
An empirical study of inter-concept similarities in multimedia ontologies
Generic concept detection has been a widely studied topic in recent research on multimedia analysis and retrieval, but the issue of how to exploit the structure of a multimedia ontology as well as different inter-concept relations, has not received similar attention. In this paper, we present results from our empirical analysis of different types of similarity among semantic concepts in two multimedia ontologies, LSCOM-Lite and CDVP-206. The results show promise that the proposed methods may be helpful in providing insight into the existing inter-concept relations within an ontology and selecting the most facilitating set of concepts and hierarchical relations. Such an analysis as this can be utilized in various tasks such as building more reliable concept detectors and designing large-scale ontologies
Probabilistic Clustering of Time-Evolving Distance Data
We present a novel probabilistic clustering model for objects that are
represented via pairwise distances and observed at different time points. The
proposed method utilizes the information given by adjacent time points to find
the underlying cluster structure and obtain a smooth cluster evolution. This
approach allows the number of objects and clusters to differ at every time
point, and no identification on the identities of the objects is needed.
Further, the model does not require the number of clusters being specified in
advance -- they are instead determined automatically using a Dirichlet process
prior. We validate our model on synthetic data showing that the proposed method
is more accurate than state-of-the-art clustering methods. Finally, we use our
dynamic clustering model to analyze and illustrate the evolution of brain
cancer patients over time
- …