884 research outputs found

    Geometric data for testing implementations of point reduction algorithms : case study using Mapshaper v 0.2.28 and previous versions

    Get PDF
    There are several open source and commercial implementations of the Visvalingam algorithm for line generalisation. The algorithm provides scope for implementation-specific interpretations, with different outcomes. This is inevitable and sometimes necessary and, they do not imply that an implementation is flawed. The only restriction is that the output must not be so inconsistent with the intent of the algorithm that it becomes inappropriate. The aim of this paper is to place the algorithm within the literature, and demonstrate the value of the teragon-test for evaluating the appropriateness of implementations; Mapshaper v 0.2.28 and earlier versions are used for illustrative purposes. Data pertaining to natural features, such as coastlines, are insufficient for establishing whether deviations in output are significant. The teragon-test produced an unexpected loss of symmetry from both the Visvalingam and Douglas-Peucker options, making the tested versions unsuitable for some applications outside of cartography. This paper describes the causes, and discusses their implications. Mapshaper 0.3.17 passes the teragon test. Other developers and users should check their implementations using contrived geometric data, such as the teragon data provided in this paper, especially when the source code is not available. The teragon-test is also useful for evaluating other point reduction algorithms

    Manufacturability and Analysis of Topologically Optimized Continuous Fiber Reinforced Composites

    Get PDF
    Researchers are unlocking the potential of Continuous Fiber Reinforced Composites for producing components with greater strength-to-weight ratios than state of the art metal alloys and unidirectional composites. The key is the emerging technology of topology optimization and advances in additive manufacturing. Topology optimization can fine tune component geometry and fiber placement all while satisfying stress constraints. However, the technology cannot yet robustly guarantee manufacturability. For this reason, substantial post-processing of an optimized design consisting of manual fiber replacement and subsequent Finite Element Analysis (FEA) is still required. To automate this post-processing in two dimensions, two (2) algorithms were developed. The first one is aimed at filling the space of a topologically optimized component with fibers of prescribed thickness. The objective is to produce flawless fiber paths, meaning no self-intersections, no tight turns, and no overlapping between fibers. It does so by leveraging concepts from elementary geometry and the Signed Distance Function of a topologically optimized domain. The manufacturable fiber paths are represented using Non-Uniform Rational Basis Splines, which can be readily conveyed to a 3D-printer as The second algorithm then calls a meshing routine to spatially discretize the topologically optimized domain. It takes input from the first algorithm to automatically create and append, orientations and material flags to the spatial elements produced by the meshing routine. Finally, it generates output that is then input to FEA software. The software is written in the C-programming language using the PETSc library. A load case is validated against MSC NASTRAN

    The Visvalingam algorithm: metrics, measures and heuristics

    Get PDF
    This paper provides the background necessary for a clear understanding of forthcoming papers relating to the Visvalingam algorithm for line generalisation, for example on the testing and usage of its implementations. It distinguishes the algorithm from implementation-specific issues to explain why it is possible to get inconsistent but equally valid output from different implementations. By tracing relevant developments within the now-disbanded Cartographic Information Systems Research Group (CISRG) of the University of Hull, it explains why a) a partial metric-driven implementation was, and still is, sufficient for many projects but not for others; b) why the Effective Area (EA) is a measure derived from a metric; c) why this measure (EA) may serve as a heuristic indicator for in-line feature segmentation and model-based generalisation; and, d) how metrics may be combined to change the order of point elimination. The issues discussed in this paper also apply to the use of other metrics. It is hoped that the background and guidance provided in this paper will enable others to participate in further research based on the algorithm

    Advances in CAD/CAM Technologies

    Get PDF

    Probabilistic lane estimation for autonomous driving using basis curves

    Get PDF
    Lane estimation for autonomous driving can be formulated as a curve estimation problem, where local sensor data provides partial and noisy observations of spatial curves forming lane boundaries. The number of lanes to estimate are initially unknown and many observations may be outliers or false detections (due e.g. to shadows or non-boundary road paint). The challenges lie in detecting lanes when and where they exist, and updating lane estimates as new observations are made. This paper describes an efficient probabilistic lane estimation algorithm based on a novel curve representation. The key advance is a principled mechanism to describe many similar curves as variations of a single basis curve. Locally observed road paint and curb features are then fused to detect and estimate all nearby travel lanes. The system handles roads with complex multi-lane geometries and makes no assumptions about the position and orientation of the vehicle with respect to the roadway. We evaluate our algorithm using a ground truth dataset containing manually-labeled, fine-grained lane geometries for vehicle travel in two large and diverse datasets that include more than 300,000 images and 44 km of roadway. The results illustrate the capabilities of our algorithm for robust lane estimation in the face of challenging conditions and unknown roadways.United States. Defense Advanced Research Projects Agency (Urban Challenge, ARPA Order No. W369/00, Program Code DIRO, issued by DARPA/CMO under Contract No. HR0011-06-C-0149

    Automatic compensating cleanup operation

    Get PDF
    Journal ArticleToday's part geometries are becoming ever more complex and require more accurate tool path to manufacture. Machining process efficiency is also a major consideration for designers as well as manufacturing engineers. Although the current advanced CAD/CAM systems have greatly improved the efficiency and accuracy of machining with the introduction of Numerically Controlled (NC) machining, excessive material may still be left on the finished part due to machining constraints, including the inaccessibility of the designed part geometry with respect the cutter, machine motion constraints like ramp angles, specific cutting patterns, etc. Polishing operations such as grinding and hand finishing are quite time consuming and expensive and may damage the surface of the part or introduce inaccuracies because of human errors. Although most of the existing machining approaches attempt to reduce such excessive restmaterials by modifying NC tool paths, none of them is satisfactory. They can be time consuming, error prone, computationally intensive, too complicated to implement, and limited to certain problem domains. A compensating cleanup tool path will be developed in this research to automatically remove these excessive material from the finish part. This method greatly reduces the burden of hand finishing and polishing and also reduces the error and complexities introduced in manually generating cleanup tool paths in the shop floor. More important, the tool path generated by this method will reduce the machining time and increase tool life compared with optimized tool path which left no excessive material behind
    • …
    corecore