10,459 research outputs found

    Strategic Roadmaps and Implementation Actions for ICT in Construction

    Get PDF

    Audiovisual preservation strategies, data models and value-chains

    No full text
    This is a report on preservation strategies, models and value-chains for digital file-based audiovisual content. The report includes: (a)current and emerging value-chains and business-models for audiovisual preservation;(b) a comparison of preservation strategies for audiovisual content including their strengths and weaknesses, and(c) a review of current preservation metadata models, and requirements for extension to support audiovisual files

    System-on-chip Computing and Interconnection Architectures for Telecommunications and Signal Processing

    Get PDF
    This dissertation proposes novel architectures and design techniques targeting SoC building blocks for telecommunications and signal processing applications. Hardware implementation of Low-Density Parity-Check decoders is approached at both the algorithmic and the architecture level. Low-Density Parity-Check codes are a promising coding scheme for future communication standards due to their outstanding error correction performance. This work proposes a methodology for analyzing effects of finite precision arithmetic on error correction performance and hardware complexity. The methodology is throughout employed for co-designing the decoder. First, a low-complexity check node based on the P-output decoding principle is designed and characterized on a CMOS standard-cells library. Results demonstrate implementation loss below 0.2 dB down to BER of 10^{-8} and a saving in complexity up to 59% with respect to other works in recent literature. High-throughput and low-latency issues are addressed with modified single-phase decoding schedules. A new "memory-aware" schedule is proposed requiring down to 20% of memory with respect to the traditional two-phase flooding decoding. Additionally, throughput is doubled and logic complexity reduced of 12%. These advantages are traded-off with error correction performance, thus making the solution attractive only for long codes, as those adopted in the DVB-S2 standard. The "layered decoding" principle is extended to those codes not specifically conceived for this technique. Proposed architectures exhibit complexity savings in the order of 40% for both area and power consumption figures, while implementation loss is smaller than 0.05 dB. Most modern communication standards employ Orthogonal Frequency Division Multiplexing as part of their physical layer. The core of OFDM is the Fast Fourier Transform and its inverse in charge of symbols (de)modulation. Requirements on throughput and energy efficiency call for FFT hardware implementation, while ubiquity of FFT suggests the design of parametric, re-configurable and re-usable IP hardware macrocells. In this context, this thesis describes an FFT/IFFT core compiler particularly suited for implementation of OFDM communication systems. The tool employs an accuracy-driven configuration engine which automatically profiles the internal arithmetic and generates a core with minimum operands bit-width and thus minimum circuit complexity. The engine performs a closed-loop optimization over three different internal arithmetic models (fixed-point, block floating-point and convergent block floating-point) using the numerical accuracy budget given by the user as a reference point. The flexibility and re-usability of the proposed macrocell are illustrated through several case studies which encompass all current state-of-the-art OFDM communications standards (WLAN, WMAN, xDSL, DVB-T/H, DAB and UWB). Implementations results are presented for two deep sub-micron standard-cells libraries (65 and 90 nm) and commercially available FPGA devices. Compared with other FFT core compilers, the proposed environment produces macrocells with lower circuit complexity and same system level performance (throughput, transform size and numerical accuracy). The final part of this dissertation focuses on the Network-on-Chip design paradigm whose goal is building scalable communication infrastructures connecting hundreds of core. A low-complexity link architecture for mesochronous on-chip communication is discussed. The link enables skew constraint looseness in the clock tree synthesis, frequency speed-up, power consumption reduction and faster back-end turnarounds. The proposed architecture reaches a maximum clock frequency of 1 GHz on 65 nm low-leakage CMOS standard-cells library. In a complex test case with a full-blown NoC infrastructure, the link overhead is only 3% of chip area and 0.5% of leakage power consumption. Finally, a new methodology, named metacoding, is proposed. Metacoding generates correct-by-construction technology independent RTL codebases for NoC building blocks. The RTL coding phase is abstracted and modeled with an Object Oriented framework, integrated within a commercial tool for IP packaging (Synopsys CoreTools suite). Compared with traditional coding styles based on pre-processor directives, metacoding produces 65% smaller codebases and reduces the configurations to verify up to three orders of magnitude

    Envisioning Digital Europe 2030: Scenarios for ICT in Future Governance and Policy Modelling

    Get PDF
    The report Envisioning Digital Europe 2030 is the result of research conducted by the Information Society Unit of IPTS as part of the CROSSROAD Project - A Participative Roadmap on ICT research on Electronic Governance and Policy Modelling (www.crossroad-eu.net ). After outlining the purpose and scope of the report and the methodological approach followed, the report presents the results of a systematic analysis of societal, policy and research trends in the governance and policy modelling domain in Europe. These analyses are considered central for understanding and roadmapping future research on ICT for governance and policy modelling. The study further illustrates the scenario design framework, analysing current and future challenges in ICT for governance and policy modelling, and identifying the key impact dimensions to be considered. It then presents the scenarios developed at the horizon 2030, including the illustrative storyboards representative of each scenario and the prospective opportunities and risks identified for each of them. The scenarios developed are internally consistent views of what the European governance and policy making system could have become by 2030 and of what the resulting implications for citizens, business and public services would be. Finally, the report draws conclusions and presents the proposed shared vision for Digital Europe 2030, offering also a summary of the main elements to be considered as an input for the future development of the research roadmap on ICT for governance and policy modelling.JRC.DDG.J.4-Information Societ

    Transforming pre-service teacher curriculum: observation through a TPACK lens

    Get PDF
    This paper will discuss an international online collaborative learning experience through the lens of the Technological Pedagogical Content Knowledge (TPACK) framework. The teacher knowledge required to effectively provide transformative learning experiences for 21st century learners in a digital world is complex, situated and changing. The discussion looks beyond the opportunity for knowledge development of content, pedagogy and technology as components of TPACK towards the interaction between those three components. Implications for practice are also discussed. In today’s technology infused classrooms it is within the realms of teacher educators, practising teaching and pre-service teachers explore and address effective practices using technology to enhance learning

    Teaching and learning in virtual worlds: is it worth the effort?

    Get PDF
    Educators have been quick to spot the enormous potential afforded by virtual worlds for situated and authentic learning, practising tasks with potentially serious consequences in the real world and for bringing geographically dispersed faculty and students together in the same space (Gee, 2007; Johnson and Levine, 2008). Though this potential has largely been realised, it generally isn’t without cost in terms of lack of institutional buy-in, steep learning curves for all participants, and lack of a sound theoretical framework to support learning activities (Campbell, 2009; Cheal, 2007; Kluge & Riley, 2008). This symposium will explore the affordances and issues associated with teaching and learning in virtual worlds, all the time considering the question: is it worth the effort

    The OCareCloudS project: toward organizing care through trusted cloud services

    Get PDF
    The increasing elderly population and the shift from acute to chronic illness makes it difficult to care for people in hospitals and rest homes. Moreover, elderly people, if given a choice, want to stay at home as long as possible. In this article, the methodologies to develop a cloud-based semantic system, offering valuable information and knowledge-based services, are presented. The information and services are related to the different personal living hemispheres of the patient, namely the daily care-related needs, the social needs and the daily life assistance. Ontologies are used to facilitate the integration, analysis, aggregation and efficient use of all the available data in the cloud. By using an interdisciplinary research approach, where user researchers, (ontology) engineers, researchers and domain stakeholders are at the forefront, a platform can be developed of great added value for the patients that want to grow old in their own home and for their caregivers
    • 

    corecore