33,194 research outputs found

    The MMT API: A Generic MKM System

    Full text link
    The MMT language has been developed as a scalable representation and interchange language for formal mathematical knowledge. It permits natural representations of the syntax and semantics of virtually all declarative languages while making MMT-based MKM services easy to implement. It is foundationally unconstrained and can be instantiated with specific formal languages. The MMT API implements the MMT language along with multiple backends for persistent storage and frontends for machine and user access. Moreover, it implements a wide variety of MMT-based knowledge management services. The API and all services are generic and can be applied to any language represented in MMT. A plugin interface permits injecting syntactic and semantic idiosyncrasies of individual formal languages.Comment: Conferences on Intelligent Computer Mathematics (CICM) 2013 The final publication is available at http://link.springer.com

    The aceToolbox: low-level audiovisual feature extraction for retrieval and classification

    Get PDF
    In this paper we present an overview of a software platform that has been developed within the aceMedia project, termed the aceToolbox, that provides global and local lowlevel feature extraction from audio-visual content. The toolbox is based on the MPEG-7 eXperimental Model (XM), with extensions to provide descriptor extraction from arbitrarily shaped image segments, thereby supporting local descriptors reflecting real image content. We describe the architecture of the toolbox as well as providing an overview of the descriptors supported to date. We also briefly describe the segmentation algorithm provided. We then demonstrate the usefulness of the toolbox in the context of two different content processing scenarios: similarity-based retrieval in large collections and scene-level classification of still images

    SDDV: scalable data dissemination in vehicular ad hoc networks

    Get PDF
    An important challenge in the domain of vehicular ad hoc networks (VANET) is the scalability of data dissemination. Under dense traffic conditions, the large number of communicating vehicles can easily result in a congested wireless channel. In that situation, delays and packet losses increase to a level where the VANET cannot be applied for road safety applications anymore. This paper introduces scalable data dissemination in vehicular ad hoc networks (SDDV), a holistic solution to this problem. It is composed of several techniques spread across the different layers of the protocol stack. Simulation results are presented that illustrate the severity of the scalability problem when applying common state-of-the-art techniques and parameters. Starting from such a baseline solution, optimization techniques are gradually added to SDDV until the scalability problem is entirely solved. Besides the performance evaluation based on simulations, the paper ends with an evaluation of the final SDDV configuration on real hardware. Experiments including 110 nodes are performed on the iMinds w-iLab.t wireless lab. The results of these experiments confirm the results obtained in the corresponding simulations

    21st Century Simulation: Exploiting High Performance Computing and Data Analysis

    Get PDF
    This paper identifies, defines, and analyzes the limitations imposed on Modeling and Simulation by outmoded paradigms in computer utilization and data analysis. The authors then discuss two emerging capabilities to overcome these limitations: High Performance Parallel Computing and Advanced Data Analysis. First, parallel computing, in supercomputers and Linux clusters, has proven effective by providing users an advantage in computing power. This has been characterized as a ten-year lead over the use of single-processor computers. Second, advanced data analysis techniques are both necessitated and enabled by this leap in computing power. JFCOM's JESPP project is one of the few simulation initiatives to effectively embrace these concepts. The challenges facing the defense analyst today have grown to include the need to consider operations among non-combatant populations, to focus on impacts to civilian infrastructure, to differentiate combatants from non-combatants, and to understand non-linear, asymmetric warfare. These requirements stretch both current computational techniques and data analysis methodologies. In this paper, documented examples and potential solutions will be advanced. The authors discuss the paths to successful implementation based on their experience. Reviewed technologies include parallel computing, cluster computing, grid computing, data logging, OpsResearch, database advances, data mining, evolutionary computing, genetic algorithms, and Monte Carlo sensitivity analyses. The modeling and simulation community has significant potential to provide more opportunities for training and analysis. Simulations must include increasingly sophisticated environments, better emulations of foes, and more realistic civilian populations. Overcoming the implementation challenges will produce dramatically better insights, for trainees and analysts. High Performance Parallel Computing and Advanced Data Analysis promise increased understanding of future vulnerabilities to help avoid unneeded mission failures and unacceptable personnel losses. The authors set forth road maps for rapid prototyping and adoption of advanced capabilities. They discuss the beneficial impact of embracing these technologies, as well as risk mitigation required to ensure success

    Knowledge is at the Edge! How to Search in Distributed Machine Learning Models

    Full text link
    With the advent of the Internet of Things and Industry 4.0 an enormous amount of data is produced at the edge of the network. Due to a lack of computing power, this data is currently send to the cloud where centralized machine learning models are trained to derive higher level knowledge. With the recent development of specialized machine learning hardware for mobile devices, a new era of distributed learning is about to begin that raises a new research question: How can we search in distributed machine learning models? Machine learning at the edge of the network has many benefits, such as low-latency inference and increased privacy. Such distributed machine learning models can also learn personalized for a human user, a specific context, or application scenario. As training data stays on the devices, control over possibly sensitive data is preserved as it is not shared with a third party. This new form of distributed learning leads to the partitioning of knowledge between many devices which makes access difficult. In this paper we tackle the problem of finding specific knowledge by forwarding a search request (query) to a device that can answer it best. To that end, we use a entropy based quality metric that takes the context of a query and the learning quality of a device into account. We show that our forwarding strategy can achieve over 95% accuracy in a urban mobility scenario where we use data from 30 000 people commuting in the city of Trento, Italy.Comment: Published in CoopIS 201

    Performance Evaluation of IPTV over WiMAX Networks Under Different Terrain Environments

    Full text link
    Deployment Video on Demand (VoD) over the next generation (WiMAX) has become one of the intense interest subjects in the research these days, and is expected to be the main revenue generators in the near future. In this paper, the performance of Quilty of Service of video streaming (IPTV) over fixed mobile WiMax network is investigated under different terrain environments, namely Free Space, Outdoor to Indoor and Pedestrian. OPNET is used to investigate the performance of VoD over WiMAX. Our findings analyzing different network statistics such as packet lost, path loss, delay, network throughput.Comment: arXiv admin note: substantial text overlap with arXiv:1302.1409, and substantial text overlap with other internet sources by other author
    corecore