174 research outputs found

    A methodology for constructing compact Chinese font libraries by radical composition.

    Get PDF
    by Wai-Yip Tung.Thesis (M.Phil.)--Chinese University of Hong Kong, 1993.Includes bibliographical references (leaves 55-56).Chapter 1. --- Introduction --- p.1Chapter 1.1. --- Previous work --- p.2Chapter 1.1.1. --- A Chinese METAFONT --- p.2Chapter 1.1.2. --- Chinese character generator --- p.2Chapter 1.1.3. --- Chinese Character Design System CCDS --- p.2Chapter 1.2. --- Goals of the thesis --- p.3Chapter 1.3. --- Overview of the thesis --- p.3Chapter 2. --- Construction of Chinese Characters --- p.5Chapter 2.1 --- Introduction --- p.5Chapter 2.2. --- liu shu(六書)Six Principles of Chinese Character Construction --- p.5Chapter 2.3. --- Structural Analysis of Chinese Characters --- p.7Chapter 2.3.1. --- Left-Right Structure --- p.8Chapter 2.3.2. --- Top-Bottom Structure --- p.9Chapter 2.3.3. --- Inside-Outside Structure --- p.10Chapter 2.3.4. --- Singleton Structure --- p.10Chapter 2.4. --- Usage frequency of radicals --- p.11Chapter 2.5. --- Usage frequency of Bushou --- p.11Chapter 2.6. --- Usage frequency of Shengpang --- p.13Chapter 2.7. --- Summary --- p.15Chapter 3. --- Composition by Radicals --- p.17Chapter 3.1. --- Introduction --- p.17Chapter 3.2. --- Transforming radicals --- p.18Chapter 3.3. --- Quality of transformed radicals --- p.19Chapter 3.4. --- Lower level components --- p.20Chapter 3.5. --- Summary --- p.23Chapter 4. --- Automatic Hinting for Chinese Font --- p.24Chapter 4.1 --- Introduction --- p.24Chapter 4.2. --- Automatic hinting for Chinese font --- p.26Chapter 4.3. --- Stroke recognition --- p.30Chapter 4.3.1. --- Identify horizontal lines --- p.31Chapter 4.3.2. --- Identify stroke segments --- p.31Chapter 4.3.3. --- Stroke recognition --- p.32Chapter 4.4. --- Regularize stroke width --- p.33Chapter 4.5. --- Grid-fitting horizontal and vertical strokes --- p.33Chapter 4.6. --- Grid-fitting radicals --- p.37Chapter 4.7. --- Summary --- p.39Chapter 5. --- RADIT - A Chinese Font Editor --- p.41Chapter 5.1. --- Introduction --- p.41Chapter 5.2. --- RADIT basics --- p.41Chapter 5.2.1. --- Character selection window --- p.42Chapter 5.2.2. --- Character window --- p.42Chapter 5.2.3. --- Tools Palette --- p.43Chapter 5.2.4. --- Toolbar --- p.43Chapter 5.2.5. --- Zooming the character window --- p.44Chapter 5.3. --- Editing a character --- p.44Chapter 5.3.1. --- Selecting handles --- p.44Chapter 5.3.2. --- Adding lines and curves --- p.45Chapter 5.3.3. --- Delete control points --- p.45Chapter 5.3.4. --- Moving control points --- p.45Chapter 5.3.5. --- Cut and paste --- p.46Chapter 5.3.6. --- Undo --- p.46Chapter 5.4. --- Adding radicals to a character --- p.46Chapter 5.5. --- Rasterizing and grid-fitting a character --- p.47Chapter 5.5.1. --- Rasterizing a character --- p.48Chapter 5.5.2. --- Stroke detection and regularization --- p.48Chapter 5.5.3. --- Grid-fitting and rasterizing a character --- p.49Chapter 6. --- Conclusions --- p.50Appendix A: Sample Fonts --- p.52References --- p.5

    Low-latency, query-driven analytics over voluminous multidimensional, spatiotemporal datasets

    Get PDF
    2017 Summer.Includes bibliographical references.Ubiquitous data collection from sources such as remote sensing equipment, networked observational devices, location-based services, and sales tracking has led to the accumulation of voluminous datasets; IDC projects that by 2020 we will generate 40 zettabytes of data per year, while Gartner and ABI estimate 20-35 billion new devices will be connected to the Internet in the same time frame. The storage and processing requirements of these datasets far exceed the capabilities of modern computing hardware, which has led to the development of distributed storage frameworks that can scale out by assimilating more computing resources as necessary. While challenging in its own right, storing and managing voluminous datasets is only the precursor to a broader field of study: extracting knowledge, insights, and relationships from the underlying datasets. The basic building block of this knowledge discovery process is analytic queries, encompassing both query instrumentation and evaluation. This dissertation is centered around query-driven exploratory and predictive analytics over voluminous, multidimensional datasets. Both of these types of analysis represent a higher-level abstraction over classical query models; rather than indexing every discrete value for subsequent retrieval, our framework autonomously learns the relationships and interactions between dimensions in the dataset (including time series and geospatial aspects), and makes the information readily available to users. This functionality includes statistical synopses, correlation analysis, hypothesis testing, probabilistic structures, and predictive models that not only enable the discovery of nuanced relationships between dimensions, but also allow future events and trends to be predicted. This requires specialized data structures and partitioning algorithms, along with adaptive reductions in the search space and management of the inherent trade-off between timeliness and accuracy. The algorithms presented in this dissertation were evaluated empirically on real-world geospatial time-series datasets in a production environment, and are broadly applicable across other storage frameworks

    Cary collection web presentation & digital image database

    Get PDF
    The Melbert Cary, Jr. Graphic Arts Collection in the Wallace Memorial Library is one of the treasures the RIT community has had available for years. For this thesis project, selected materials from the collection were scanned and made available on the Internet so people at any location can experience the rare and invaluable items the facility houses. Not only has the result of this project created an educational tool for others to use, but it also challenged the author to master web publishing while developments rapidly occur on the most powerful mass communications media to arise in decades. While the primary purpose of this thesis project was to create an aesthetically pleasing and information rich web presentation for the Melbert Cary, Jr. Graphic Arts Collection, many secondary goals had to first be achieved. Those secondary goals are outlined in this list: 1. To acquire high quality color electronic images for others to access remotely. 2. To design a searchable database of 300 records. 3. To learn the ins-and-outs of web publishing by: Creating cohesive and consistent documents in the HyperText Markup Language (HTML). Developing an aesthetically pleasing interface for users to explore documents. This included keeping up-to-date with developments in HTML and using techniques created by web publishing experts to make the text as typographically pleasing as possible. Placing the necessary documents and images on a web server. Advertising the address of the presentation, or URL, to the appropriate audience. 4. Developing clear and concise instructions on how to maintain the presentation, including procedures for adding categories and images. After a substantial amount of work on this project was completed, it was linked to the Wallace Memorial Library\u27s home page. A home page is the first site a person reaches upon typing in an Internet address on the World Wide Web. Home pages can be created by individuals or organizations and serve as points of departure for exploring textual and graphical information available at these sites. The Cary presentation has a section explaining the history and growth of the collection. A user can continue by taking a virtual tour of the facilities or by reading about recent acquisitions. The main feature of the presentation is a subject library and digital image data base which contains an initial collection of approximately 300 searchable records ranging from medieval manuscripts to portraits of printers. Instructions on how to perform various types of searches, as well as what type of searches are feasible, are also integrated into this project. Finally, information about image acquisition, graphics presentation, and database installation and setup are integrated into this project. In addition, a secondary home page for the American Printing History Association has been created, and sample articles from its journal Printing History will be available in a digital format on an ongoing basis. All images were prepared to be as faithful to the originals as possible, keeping in mind the drawbacks inherent in viewing images and text on today\u27s monitors. Retaining accurate colors and details while paying heed to practical speed requirements of transmission was of great importance. A feedback form has also been made available for individuals who wish to communicate any comments, problems, requests or suggestions via e-mail. The documentation that follows is the methodology used in creating the Cary Collection\u27s web presentation

    Legibility of condensed perceptually-tuned grayscale fonts

    Get PDF
    The authors analyze the quality of condensed text on LCD displays, generated with unhinted and hinted bilevel characters, with traditional anti-aliased and with perceptually-tuned grayscale characters. Hinted bi-level characters and perceptually-tuned grayscale characters improve the quality of displayed small size characters (8pt, 6pt) up to a line condensation factor of 80%. At higher condensation factors, the text becomes partly illegible. In such situations, traditional anti-aliased grayscale character seems to be the most robust variant. They explore the utility of perceptually-tuned grayscale fonts for improving the legibility of condensed text. A small advantage was found for text searching, compared to bilevel fonts. This advantage is consistent with human vision models applied to readin

    Weiterentwicklung analytischer Datenbanksysteme

    Get PDF
    This thesis contributes to the state of the art in analytical database systems. First, we identify and explore extensions to better support analytics on event streams. Second, we propose a novel polygon index to enable efficient geospatial data processing in main memory. Third, we contribute a new deep learning approach to cardinality estimation, which is the core problem in cost-based query optimization.Diese Arbeit trägt zum aktuellen Forschungsstand von analytischen Datenbanksystemen bei. Wir identifizieren und explorieren Erweiterungen um Analysen auf Eventströmen besser zu unterstützen. Wir stellen eine neue Indexstruktur für Polygone vor, die eine effiziente Verarbeitung von Geodaten im Hauptspeicher ermöglicht. Zudem präsentieren wir einen neuen Ansatz für Kardinalitätsschätzungen mittels maschinellen Lernens

    Searching for patterns in Conway's Game of Life

    Get PDF
    Conway’s Game of Life (Life) is a simple cellular automaton, discovered by John Conway in 1970, that exhibits complex emergent behavior. Life-enthusiasts have been looking for building blocks with specific properties (patterns) to answer unsolved problems in Life for the past five decades. Finding patterns in Life is difficult due to the large search space. Current search algorithms use an explorative approach based on the rules of the game, but this can only sample a small fraction of the search space. More recently, people have used Sat solvers to search for patterns. These solvers are not specifically tuned to this problem and thus waste a lot of time processing Life’s rules in an engine that does not understand them. We propose a novel Sat-based approach that replaces the binary tree used by traditional Sat solvers with a grid-based approach, complemented by an injection of Game of Life specific knowledge. This leads to a significant speedup in searching. As a fortunate side effect, our solver can be generalized to solve general Sat problems. Because it is grid-based, all manipulations are embarrassingly parallel, allowing implementation on massively parallel hardware

    Efficient algorithms for passive network measurement

    Get PDF
    Network monitoring has become a necessity to aid in the management and operation of large networks. Passive network monitoring consists of extracting metrics (or any information of interest) by analyzing the traffic that traverses one or more network links. Extracting information from a high-speed network link is challenging, given the great data volumes and short packet inter-arrival times. These difficulties can be alleviated by using extremely efficient algorithms or by sampling the incoming traffic. This work improves the state of the art in both these approaches. For one-way packet delay measurement, we propose a series of improvements over a recently appeared technique called Lossy Difference Aggregator. A main limitation of this technique is that it does not provide per-flow measurements. We propose a data structure called Lossy Difference Sketch that is capable of providing such per-flow delay measurements, and, unlike recent related works, does not rely on any model of packet delays. In the problem of collecting measurements under the sliding window model, we focus on the estimation of the number of active flows and in traffic filtering. Using a common approach, we propose one algorithm for each problem that obtains great accuracy with significant resource savings. In the traffic sampling area, the selection of the sampling rate is a crucial aspect. The most sensible approach involves dynamically adjusting sampling rates according to network traffic conditions, which is known as adaptive sampling. We propose an algorithm called Cuckoo Sampling that can operate with a fixed memory budget and perform adaptive flow-wise packet sampling. It is based on a very simple data structure and is computationally extremely lightweight. The techniques presented in this work are thoroughly evaluated through a combination of theoretical and experimental analysis.Postprint (published version
    • …
    corecore