2,333 research outputs found

    Incremental file reorganization schemes

    Get PDF
    Issued as Final project report, Project no. G-36-66

    A Survey on Array Storage, Query Languages, and Systems

    Full text link
    Since scientific investigation is one of the most important providers of massive amounts of ordered data, there is a renewed interest in array data processing in the context of Big Data. To the best of our knowledge, a unified resource that summarizes and analyzes array processing research over its long existence is currently missing. In this survey, we provide a guide for past, present, and future research in array processing. The survey is organized along three main topics. Array storage discusses all the aspects related to array partitioning into chunks. The identification of a reduced set of array operators to form the foundation for an array query language is analyzed across multiple such proposals. Lastly, we survey real systems for array processing. The result is a thorough survey on array data storage and processing that should be consulted by anyone interested in this research topic, independent of experience level. The survey is not complete though. We greatly appreciate pointers towards any work we might have forgotten to mention.Comment: 44 page

    Concurrent Context-Free Framework for Conceptual Similarity Problem using Reverse Dictionary

    Get PDF
    Semantic search is one of the most prominent options to search the required and relevant content from the web. But most of them are doing key word and phrase wise similarity search. It may or may not find the relevant information because they directly search with that phrase. But, in most of the cases documents may conceptually equal instead of term wise. Reverse dictionary can solve such type of problems. This will take meaning of the word and it will return related keywords with respective ranks. But main problem here is building such dictionaries is time and memory consuming. Cost effective solutions are required to reduce search time and in-memory requirements. This paper focuses on such aspects by utilizing concurrent programming and efficient index structures and builds a framework to Conceptual similarity problem using reverse dictionary. Simulation results shows that proposed approach can take less time when compared to existing approaches

    STANSE: Bug-finding Framework for C Programs

    Get PDF
    Regular paper accepted at the MEMICS 2011 workshop. The paper deals with static analysis. It also describes a framework and tool called Stanse

    HPTA: High-Performance Text Analytics

    Get PDF

    Data-centric concurrency control on the java programming language

    Get PDF
    Dissertação para obtenção do Grau de Mestre em Engenharia InformáticaThe multi-core paradigm has propelled shared-memory concurrent programming to an important role in software development. Its use is however limited by the constructs that provide a layer of abstraction for synchronizing access to shared resources. Reasoning with these constructs is not trivial due to their concurrent nature. Data-races and deadlocks occur in concurrent programs, encumbering the programmer and further reducing his productivity. Even though the constructs should be as unobtrusive and intuitive as possible, performance must also be kept high compared to legacy lock-based mechanism. Failure to guarantee similar performance will hinder a system from adoption. Recent research attempts to address these issues. However, the current state of the art in concurrency control mechanisms is mostly code-centric and not intuitive. Its codecentric nature requires the specification of the zones in the code that require synchronization,contributing to the decentralization of concurrency bugs and error-proneness of the programmer. On the other hand, the only data-centric approach, AJ [VTD06], exposes excessive detail to the programmer and fails to provide complete deadlock-freedom. Given this state of the art, our proposal intends to provide the programmer a set of unobtrusive data-centric constructs. These will guarantee desirable security properties: composability, atomicity, and deadlock-freedom in all scenarios. For that purpose, a lower level mechanism (ResourceGroups) will be used. The model proposed resides on the known concept of atomic variables, the basis for our concurrency control mechanism. To infer the efficiency of our work, it is compared to Java synchronized blocks, transactional memory and AJ, where our system demonstrates a competitive performance and an equivalent level of expressivity.RepComp project(PTDC/EIA-EIA/108963/2008

    The Table Generating Routines of a Data Description Language Processor

    Get PDF
    The Data Description Language Processor, designed by J. A. Ramirez, is the compiler for a modified version of the Data Description Language (DDL), written by D. P. Smith. Two main phases exist in the DDL Processor: 1) The Syntactic Analysis phase and 2) The Code Generation phase The former phase checks the DDL source for local and global syntactic flaws before passing control to the latter. In order to speed up execution of phase 2, internal tables (one symbol and several data tables), containing encoded versions of the DDL source input, are constructed. The tables, created during syntax analysis, will facilitate global syntax checking (verifying all DDL statement references to be valid), and will permit code generation to operate more quickly by providing it with the essence of the source data and, hence, negate the necessity of a second pass over the source input

    Principles of Information Systems Analysis and Design

    Get PDF
    • …
    corecore