12,178 research outputs found

    3D Visibility Representations of 1-planar Graphs

    Full text link
    We prove that every 1-planar graph G has a z-parallel visibility representation, i.e., a 3D visibility representation in which the vertices are isothetic disjoint rectangles parallel to the xy-plane, and the edges are unobstructed z-parallel visibilities between pairs of rectangles. In addition, the constructed representation is such that there is a plane that intersects all the rectangles, and this intersection defines a bar 1-visibility representation of G.Comment: Appears in the Proceedings of the 25th International Symposium on Graph Drawing and Network Visualization (GD 2017

    Visibility Representations of Boxes in 2.5 Dimensions

    Full text link
    We initiate the study of 2.5D box visibility representations (2.5D-BR) where vertices are mapped to 3D boxes having the bottom face in the plane z=0z=0 and edges are unobstructed lines of sight parallel to the xx- or yy-axis. We prove that: (i)(i) Every complete bipartite graph admits a 2.5D-BR; (ii)(ii) The complete graph KnK_n admits a 2.5D-BR if and only if n19n \leq 19; (iii)(iii) Every graph with pathwidth at most 77 admits a 2.5D-BR, which can be computed in linear time. We then turn our attention to 2.5D grid box representations (2.5D-GBR) which are 2.5D-BRs such that the bottom face of every box is a unit square at integer coordinates. We show that an nn-vertex graph that admits a 2.5D-GBR has at most 4n6n4n - 6 \sqrt{n} edges and this bound is tight. Finally, we prove that deciding whether a given graph GG admits a 2.5D-GBR with a given footprint is NP-complete. The footprint of a 2.5D-BR Γ\Gamma is the set of bottom faces of the boxes in Γ\Gamma.Comment: Appears in the Proceedings of the 24th International Symposium on Graph Drawing and Network Visualization (GD 2016

    Pixel and Voxel Representations of Graphs

    Full text link
    We study contact representations for graphs, which we call pixel representations in 2D and voxel representations in 3D. Our representations are based on the unit square grid whose cells we call pixels in 2D and voxels in 3D. Two pixels are adjacent if they share an edge, two voxels if they share a face. We call a connected set of pixels or voxels a blob. Given a graph, we represent its vertices by disjoint blobs such that two blobs contain adjacent pixels or voxels if and only if the corresponding vertices are adjacent. We are interested in the size of a representation, which is the number of pixels or voxels it consists of. We first show that finding minimum-size representations is NP-complete. Then, we bound representation sizes needed for certain graph classes. In 2D, we show that, for kk-outerplanar graphs with nn vertices, Θ(kn)\Theta(kn) pixels are always sufficient and sometimes necessary. In particular, outerplanar graphs can be represented with a linear number of pixels, whereas general planar graphs sometimes need a quadratic number. In 3D, Θ(n2)\Theta(n^2) voxels are always sufficient and sometimes necessary for any nn-vertex graph. We improve this bound to Θ(nτ)\Theta(n\cdot \tau) for graphs of treewidth τ\tau and to O((g+1)2nlog2n)O((g+1)^2n\log^2n) for graphs of genus gg. In particular, planar graphs admit representations with O(nlog2n)O(n\log^2n) voxels

    Virtual Rephotography: Novel View Prediction Error for 3D Reconstruction

    Full text link
    The ultimate goal of many image-based modeling systems is to render photo-realistic novel views of a scene without visible artifacts. Existing evaluation metrics and benchmarks focus mainly on the geometric accuracy of the reconstructed model, which is, however, a poor predictor of visual accuracy. Furthermore, using only geometric accuracy by itself does not allow evaluating systems that either lack a geometric scene representation or utilize coarse proxy geometry. Examples include light field or image-based rendering systems. We propose a unified evaluation approach based on novel view prediction error that is able to analyze the visual quality of any method that can render novel views from input images. One of the key advantages of this approach is that it does not require ground truth geometry. This dramatically simplifies the creation of test datasets and benchmarks. It also allows us to evaluate the quality of an unknown scene during the acquisition and reconstruction process, which is useful for acquisition planning. We evaluate our approach on a range of methods including standard geometry-plus-texture pipelines as well as image-based rendering techniques, compare it to existing geometry-based benchmarks, and demonstrate its utility for a range of use cases.Comment: 10 pages, 12 figures, paper was submitted to ACM Transactions on Graphics for revie
    corecore