28,987 research outputs found

    Natural Transmission of Information Extraction Results to End-Users - A Proof-of-Concept Using Data-to-Text

    Get PDF
    Information Extraction from natural texts has a great potential in areas such as Tourism and can be of great assistance in transforming customers\u27 comments in valuable information for Tourism operators, governments and customers. After extraction, information needs to be efficiently transmitted to end-users in a natural way. Systems should not, in general, send extracted information directly to end-users, such as hotel managers, as it can be difficult to read. Naturally, humans transmit and encode information using natural languages, such as Portuguese. The problem arising from the need of efficient and natural transmission of the information to end-user is how to encode it. The use of natural language generation (NLG) is a possible solution, for producing sentences, and, with them, texts. In this paper we address this, with a data-to-text system, a derivation of formal NLG systems that use data as input. The proposed system uses an aligned corpus, which was defined, collected and processed, in about approximately 3 weeks of work. To build the language model were used three different in-domain and out-of-domain corpora. The effects of this approach were evaluated, and results are presented. Automatic metrics, BLEU and Meteor, were used to evaluate the different systems, comparing their values with similar systems. Results show that expanding the corpus has a major positive effect in BLEU and Meteor scores and use of additional corpora (in-domain and out-of-domain) in training language model does not result in significantly different performance. The scores obtained, combined with their comparison with other systems performance and informal evaluation by humans of the sentences produced, give additional support for the capabilities of the translation based approach for fast development of data-to-text for new domains

    5GNOW: Challenging the LTE Design Paradigms of Orthogonality and Synchronicity

    Full text link
    LTE and LTE-Advanced have been optimized to deliver high bandwidth pipes to wireless users. The transport mechanisms have been tailored to maximize single cell performance by enforcing strict synchronism and orthogonality within a single cell and within a single contiguous frequency band. Various emerging trends reveal major shortcomings of those design criteria: 1) The fraction of machine-type-communications (MTC) is growing fast. Transmissions of this kind are suffering from the bulky procedures necessary to ensure strict synchronism. 2) Collaborative schemes have been introduced to boost capacity and coverage (CoMP), and wireless networks are becoming more and more heterogeneous following the non-uniform distribution of users. Tremendous efforts must be spent to collect the gains and to manage such systems under the premise of strict synchronism and orthogonality. 3) The advent of the Digital Agenda and the introduction of carrier aggregation are forcing the transmission systems to deal with fragmented spectrum. 5GNOW is an European research project supported by the European Commission within FP7 ICT Call 8. It will question the design targets of LTE and LTE-Advanced having these shortcomings in mind and the obedience to strict synchronism and orthogonality will be challenged. It will develop new PHY and MAC layer concepts being better suited to meet the upcoming needs with respect to service variety and heterogeneous transmission setups. Wireless transmission networks following the outcomes of 5GNOW will be better suited to meet the manifoldness of services, device classes and transmission setups present in envisioned future scenarios like smart cities. The integration of systems relying heavily on MTC into the communication network will be eased. The per-user experience will be more uniform and satisfying. To ensure this 5GNOW will contribute to upcoming 5G standardization.Comment: Submitted to Workshop on Mobile and Wireless Communication Systems for 2020 and beyond (at IEEE VTC 2013, Spring

    Safe to Be Open: Study on the Protection of Research Data and Recommendations for Access and Usage

    Get PDF
    Openness has become a common concept in a growing number of scientific and academic fields. Expressions such as Open Access (OA) or Open Content (OC) are often employed for publications of papers and research results, or are contained as conditions in tenders issued by a number of funding agencies. More recently the concept of Open Data (OD) is of growing interest in some fields, particularly those that produce large amounts of data – which are not usually protected by standard legal tools such as copyright. However, a thorough understanding of the meaning of Openness – especially its legal implications – is usually lacking. Open Access, Public Access, Open Content, Open Data, Public Domain. All these terms are often employed to indicate that a given paper, repository or database does not fall under the traditional “closed” scheme of default copyright rules. However, the differences between all these terms are often largely ignored or misrepresented, especially when the scientist in question is not familiar with the law generally and copyright in particular – a very common situation in all scientific fields. On 17 July 2012 the European Commission published its Communication to the European Parliament and the Council entitled “Towards better access to scientific information: Boosting the benefits of public investments in research”. As the Commission observes, “discussions of the scientific dissemination system have traditionally focused on access to scientific publications – journals and monographs. However, it is becoming increasingly important to improve access to research data (experimental results, observations and computer-generated information), which forms the basis for the quantitative analysis underpinning many scientific publications”. The Commission believes that through more complete and wider access to scientific publications and data, the pace of innovation will accelerate and researchers will collaborate so that duplication of efforts will be avoided. Moreover, open research data will allow other researchers to build on previous research results, as it will allow involvement of citizens and society in the scientific process. In the Communication the Commission makes explicit reference to open access models of publications and dissemination of research results, and the reference is not only to access and use but most significantly to reuse of publications as well as research data. The Communication marks an official new step on the road to open access to publicly funded research results in science and the humanities in Europe. Scientific publications are no longer the only elements of its open access policy: research data upon which publications are based should now also be made available to the public. As noble as the open access goal is, however, the expansion of the open access policy to publicly funded research data raises a number of legal and policy issues that are often distinct from those concerning the publication of scientific articles and monographs. Since open access to research data – rather than publications – is a relatively new policy objective, less attention has been paid to the specific features of research data. An analysis of the legal status of such data, and on how to make it available under the correct licence terms, is therefore the subject of the following sections

    Establishing effective communications in disaster affected areas and artificial intelligence based detection using social media platform

    Get PDF
    Floods, earthquakes, storm surges and other natural disasters severely affect the communication infrastructure and thus compromise the effectiveness of communications dependent rescue and warning services. In this paper, a user centric approach is proposed to establish communications in disaster affected and communication outage areas. The proposed scheme forms ad hoc clusters to facilitate emergency communications and connect end-users/ User Equipment (UE) to the core network. A novel cluster formation with single and multi-hop communication framework is proposed. The overall throughput in the formed clusters is maximized using convex optimization. In addition, an intelligent system is designed to label different clusters and their localities into affected and non-affected areas. As a proof of concept, the labeling is achieved on flooding dataset where region specific social media information is used in proposed machine learning techniques to classify the disaster-prone areas as flooded or unflooded. The suitable results of the proposed machine learning schemes suggest its use along with proposed clustering techniques to revive communications in disaster affected areas and to classify the impact of disaster for different locations in disaster-prone areas

    AI-assisted patent prior art searching - feasibility study

    Get PDF
    This study seeks to understand the feasibility, technical complexities and effectiveness of using artificial intelligence (AI) solutions to improve operational processes of registering IP rights. The Intellectual Property Office commissioned Cardiff University to undertake this research. The research was funded through the BEIS Regulators’ Pioneer Fund (RPF). The RPF fund was set up to help address barriers to innovation in the UK economy

    Protecting Voice Controlled Systems Using Sound Source Identification Based on Acoustic Cues

    Full text link
    Over the last few years, a rapidly increasing number of Internet-of-Things (IoT) systems that adopt voice as the primary user input have emerged. These systems have been shown to be vulnerable to various types of voice spoofing attacks. Existing defense techniques can usually only protect from a specific type of attack or require an additional authentication step that involves another device. Such defense strategies are either not strong enough or lower the usability of the system. Based on the fact that legitimate voice commands should only come from humans rather than a playback device, we propose a novel defense strategy that is able to detect the sound source of a voice command based on its acoustic features. The proposed defense strategy does not require any information other than the voice command itself and can protect a system from multiple types of spoofing attacks. Our proof-of-concept experiments verify the feasibility and effectiveness of this defense strategy.Comment: Proceedings of the 27th International Conference on Computer Communications and Networks (ICCCN), Hangzhou, China, July-August 2018. arXiv admin note: text overlap with arXiv:1803.0915

    AN EVOLUTIONARY APPROACH TO BIBLIOGRAPHIC CLASSIFICATION

    Get PDF
    This dissertation is research in the domain of information science and specifically, the organization and representation of information. The research has implications for classification of scientific books, especially as dissemination of information becomes more rapid and science becomes more diverse due to increases in multi-, inter-, trans-disciplinary research, which focus on phenomena, in contrast to traditional library classification schemes based on disciplines.The literature review indicates 1) human socio-cultural groups have many of the same properties as biological species, 2) output from human socio-cultural groups can be and has been the subject of evolutionary relationship analyses (i.e., phylogenetics), 3) library and information science theorists believe the most favorable and scientific classification for information packages is one based on common origin, but 4) library and information science classification researchers have not demonstrated a book classification based on evolutionary relationships of common origin.The research project supports the assertion that a sensible book classification method can be developed using a contemporary biological classification approach based on common origin, which has not been applied to a collection of books until now. Using a sample from a collection of earth-science digitized books, the method developed includes a text-mining step to extract important terms, which were converted into a dataset for input into the second step—the phylogenetic analysis. Three classification trees were produced and are discussed. Parsimony analysis, in contrast to distance and likelihood analyses, produced a sensible book classification tree. Also included is a comparison with a classification tree based on a well-known contemporary library classification scheme (the Library of Congress Classification).Final discussions connect this research with knowledge organization and information retrieval, information needs beyond science, and this type of research in context of a unified science of cultural evolution
    • …
    corecore