1,217 research outputs found
Recommended from our members
Access and comprehension - Teachers use of simplified language materials
This paper examines the current use of simplified language materials (SLMs) by primary and secondary teachers across England. Drawing on a survey of 33 schools the paper examines the degree to which teachers and support teachers currently use simplified language materials and the reasons they give for their usage. It discusses both the contradictions and similarities between teachers' perceptions of the value of SLMs and the existing research base. It focuses on current national guidance, the role of SLMs for people with learning difficulties and research that encourages the use of complex materials and bilingual support, contrasting this with teachers perception that SLMs increase both access and comprehension. The paper suggests that we should not expect teachers to abandon SLMs but should find ways to use this skill base to enhance the education of all
Speeding up liquid crystal SLMs using overdrive with phase change reduction
Nematic liquid crystal spatial light modulators (SLMs) with fast switching times and high diffraction efficiency are important to various applications ranging from optical beam steering and adaptive optics to optical tweezers. Here we demonstrate the great benefits that can be derived in terms of speed enhancement without loss of diffraction efficiency from two mutually compatible approaches. The first technique involves the idea of overdrive, that is the calculation of intermediate patterns to speed up the transition to the target phase pattern. The second concerns optimization of the target pattern to reduce the required phase change applied to each pixel, which in addition leads to a substantial reduction of variations in the intensity of the diffracted light during the transition. When these methods are applied together, we observe transition times for the diffracted light fields of about 1 ms, which represents up to a tenfold improvement over current approaches. We experimentally demonstrate the improvements of the approach for applications such as holographic image projection, beam steering and switching, and real-time control loops
The ESCAPE project : Energy-efficient Scalable Algorithms for Weather Prediction at Exascale
In the simulation of complex multi-scale flows arising in weather and climate modelling, one of the biggest challenges is to satisfy strict service requirements in terms of time to solution and to satisfy budgetary constraints in terms of energy to solution, without compromising the accuracy and stability of the application. These simulations require algorithms that minimise the energy footprint along with the time required to produce a solution, maintain the physically required level of accuracy, are numerically stable, and are resilient in case of hardware failure.
The European Centre for Medium-Range Weather Forecasts (ECMWF) led the ESCAPE (Energy-efficient Scalable Algorithms for Weather Prediction at Exascale) project, funded by Horizon 2020 (H2020) under the FET-HPC (Future and Emerging Technologies in High Performance Computing) initiative. The goal of ESCAPE was to develop a sustainable strategy to evolve weather and climate prediction models to next-generation computing technologies. The project partners incorporate the expertise of leading European regional forecasting consortia, university research, experienced high-performance computing centres, and hardware vendors.
This paper presents an overview of the ESCAPE strategy: (i) identify domain-specific key algorithmic motifs in weather prediction and climate models (which we term Weather & Climate Dwarfs), (ii) categorise them in terms of computational and communication patterns while (iii) adapting them to different hardware architectures with alternative programming models, (iv) analyse the challenges in optimising, and (v) find alternative algorithms for the same scheme. The participating weather prediction models are the following: IFS (Integrated Forecasting System); ALARO, a combination of AROME (Application de la Recherche a l'Operationnel a Meso-Echelle) and ALADIN (Aire Limitee Adaptation Dynamique Developpement International); and COSMO-EULAG, a combination of COSMO (Consortium for Small-scale Modeling) and EULAG (Eulerian and semi-Lagrangian fluid solver). For many of the weather and climate dwarfs ESCAPE provides prototype implementations on different hardware architectures (mainly Intel Skylake CPUs, NVIDIA GPUs, Intel Xeon Phi, Optalysys optical processor) with different programming models. The spectral transform dwarf represents a detailed example of the co-design cycle of an ESCAPE dwarf.
The dwarf concept has proven to be extremely useful for the rapid prototyping of alternative algorithms and their interaction with hardware; e.g. the use of a domain-specific language (DSL). Manual adaptations have led to substantial accelerations of key algorithms in numerical weather prediction (NWP) but are not a general recipe for the performance portability of complex NWP models. Existing DSLs are found to require further evolution but are promising tools for achieving the latter. Measurements of energy and time to solution suggest that a future focus needs to be on exploiting the simultaneous use of all available resources in hybrid CPU-GPU arrangements
Angular EPR paradox
The violation of local uncertainty relations is a valuable tool for detecting
entanglement, especially in multi-dimensional systems. The orbital angular
momentum of light provides such a multi-dimensional system. We study quantum
correlations for the conjugate variables of orbital angular momentum and
angular position. We determine an experimentally testable criterion for the
demonstration of an angular version of the EPR paradox. For the interpretation
of future experimental results from our proposed setup, we include a model for
the indeterminacies inherent to the angular position measurement. For this
measurement angular apertures are used to determine the probability density of
the angle. We show that for a class of aperture functions a demonstration of an
angular EPR paradox, according to our criterion, is to be expected.Comment: 21 pages, 9 figures, to be published in J. Mod. Opt. special issue on
quantum imagin
Mind's Mirror: Distilling Self-Evaluation Capability and Comprehensive Thinking from Large Language Models
Large language models (LLMs) have achieved remarkable advancements in the
field of natural language processing. However, the sheer scale and
computational demands of these models present formidable challenges when
considering their practical deployment in resource-constrained contexts. While
techniques such as chain-of-thought (CoT) distillation have displayed promise
in distilling LLMs into small language models (SLMs), there is a risk that
distilled SLMs may still carry over flawed reasoning or hallucinations
inherited from their LLM counterparts. To address these issues, we propose a
twofold methodology: First, we introduce a novel method for distilling the
self-evaluation capability inherent in LLMs into SLMs, which aims to mitigate
the adverse effects of erroneous reasoning and reduce hallucinations. Second,
we advocate for a comprehensive distillation process that incorporates multiple
distinct chain-of-thought and self-evaluation paradigms and ensures a more
holistic and robust knowledge transfer into SLMs. Experiments on three NLP
benchmarks demonstrate that our method significantly improves the performance
of distilled SLMs and sheds light on the path towards developing smaller models
closely aligned with human cognition.Comment: 13 pages, 5 figure
Learning Distributed Representations for Multiple-Viewpoint Melodic Prediction
The analysis of sequences is important for extracting in- formation from music owing to its fundamentally temporal nature. In this paper, we present a distributed model based on the Restricted Boltzmann Machine (RBM) for learning melodic sequences. The model is similar to a previous suc- cessful neural network model for natural language [2]. It is first trained to predict the next pitch in a given pitch se- quence, and then extended to also make use of information in sequences of note-durations in monophonic melodies on the same task. In doing so, we also propose an efficient way of representing this additional information that takes advantage of the RBM’s structure. Results show that this RBM-based prediction model performs better than previ- ously evaluated n-gram models and also outperforms them in certain cases. It is able to make use of information present in longer sequences more effectively than n-gram models, while scaling linearly in the number of free pa- rameters required
Scaling Down to Scale Up: A Cost-Benefit Analysis of Replacing OpenAI's LLM with Open Source SLMs in Production
Many companies use large language models (LLMs) offered as a service, like
OpenAI's GPT-4, to create AI-enabled product experiences. Along with the
benefits of ease-of-use and shortened time-to-solution, this reliance on
proprietary services has downsides in model control, performance reliability,
uptime predictability, and cost. At the same time, a flurry of open-source
small language models (SLMs) has been made available for commercial use.
However, their readiness to replace existing capabilities remains unclear, and
a systematic approach to holistically evaluate these SLMs is not readily
available. This paper presents a systematic evaluation methodology and a
characterization of modern open-source SLMs and their trade-offs when replacing
proprietary LLMs for a real-world product feature. We have designed SLaM, an
open-source automated analysis tool that enables the quantitative and
qualitative testing of product features utilizing arbitrary SLMs. Using SLaM,
we examine the quality and performance characteristics of modern SLMs relative
to an existing customer-facing implementation using the OpenAI GPT-4 API.
Across 9 SLMs and their 29 variants, we observe that SLMs provide competitive
results, significant performance consistency improvements, and a cost reduction
of 5x~29x when compared to GPT-4.Comment: Updated title, Revised conten
Small Language Models for Curriculum-based Guidance
The adoption of generative AI and large language models (LLMs) in education is still emerging. In this study, we explore the development and evaluation of AI teaching assistants that provide curriculum-based guidance using a retrieval-augmented generation (RAG) pipeline applied to selected open-source small language models (SLMs). We benchmarked eight SLMs, including LLaMA 3.1, IBM Granite 3.3, and Gemma 3 (7–17B parameters), against GPT-4o. Our findings show that with proper prompting and targeted retrieval, SLMs can match LLMs in delivering accurate, pedagogically aligned responses. Importantly, SLMs offer significant sustainability benefits due to their lower computational and energy requirements, enabling real-time use on consumer-grade hardware without depending on cloud infrastructure. This makes them not only cost-effective and privacy-preserving but also environmentally responsible, positioning them as viable AI teaching assistants for educational institutions aiming to scale personalized learning in a sustainable and energy-efficient manner
- …
