14,504 research outputs found
An exploration of the language within Ofsted reports and their influence on primary school performance in mathematics: a mixed methods critical discourse analysis
This thesis contributes to the understanding of the language of Ofsted reports, their similarity to one another and associations between different terms used within ‘areas for improvement’ sections and subsequent outcomes for pupils. The research responds to concerns from serving headteachers that Ofsted reports are overly similar, do not capture the unique story of their school, and are unhelpful for improvement. In seeking to answer ‘how similar are
Ofsted reports’ the study uses two tools, a plagiarism detection software (Turnitin) and a discourse analysis tool (NVivo) to identify trends within and across a large corpus of reports.
The approach is based on critical discourse analysis (Van Dijk, 2009; Fairclough, 1989) but shaped in the form of practitioner enquiry seeking power in the form of impact on pupils and practitioners, rather than a more traditional, sociological application of the method.
The research found that in 2017, primary school section 5 Ofsted reports had more than half of their content exactly duplicated within other primary school inspection reports published that same year. Discourse analysis showed the quality assurance process overrode variables such as inspector designation, gender, or team size, leading to three distinct patterns of duplication: block duplication, self-referencing, and template writing. The most unique part of a report was found to be the ‘area for improvement’ section, which was tracked to externally verified outcomes for pupils using terms linked to ‘mathematics’. Those
required to improve mathematics in their areas for improvement improved progress and attainment in mathematics significantly more than national rates. These findings indicate that there was a positive correlation between the inspection reporting process and a beneficial impact on pupil outcomes in mathematics, and that the significant similarity of one report to another had no bearing on the usefulness of the report for school improvement purposes
within this corpus
Self-Supervised Learning to Prove Equivalence Between Straight-Line Programs via Rewrite Rules
We target the problem of automatically synthesizing proofs of semantic
equivalence between two programs made of sequences of statements. We represent
programs using abstract syntax trees (AST), where a given set of
semantics-preserving rewrite rules can be applied on a specific AST pattern to
generate a transformed and semantically equivalent program. In our system, two
programs are equivalent if there exists a sequence of application of these
rewrite rules that leads to rewriting one program into the other. We propose a
neural network architecture based on a transformer model to generate proofs of
equivalence between program pairs. The system outputs a sequence of rewrites,
and the validity of the sequence is simply checked by verifying it can be
applied. If no valid sequence is produced by the neural network, the system
reports the programs as non-equivalent, ensuring by design no programs may be
incorrectly reported as equivalent. Our system is fully implemented for a given
grammar which can represent straight-line programs with function calls and
multiple types. To efficiently train the system to generate such sequences, we
develop an original incremental training technique, named self-supervised
sample selection. We extensively study the effectiveness of this novel training
approach on proofs of increasing complexity and length. Our system, S4Eq,
achieves 97% proof success on a curated dataset of 10,000 pairs of equivalent
programsComment: 30 pages including appendi
Examples of works to practice staccato technique in clarinet instrument
Klarnetin staccato tekniğini güçlendirme aşamaları eser çalışmalarıyla uygulanmıştır. Staccato
geçişlerini hızlandıracak ritim ve nüans çalışmalarına yer verilmiştir. Çalışmanın en önemli amacı
sadece staccato çalışması değil parmak-dilin eş zamanlı uyumunun hassasiyeti üzerinde de
durulmasıdır. Staccato çalışmalarını daha verimli hale getirmek için eser çalışmasının içinde etüt
çalışmasına da yer verilmiştir. Çalışmaların üzerinde titizlikle durulması staccato çalışmasının ilham
verici etkisi ile müzikal kimliğe yeni bir boyut kazandırmıştır. Sekiz özgün eser çalışmasının her
aşaması anlatılmıştır. Her aşamanın bir sonraki performans ve tekniği güçlendirmesi esas alınmıştır.
Bu çalışmada staccato tekniğinin hangi alanlarda kullanıldığı, nasıl sonuçlar elde edildiği bilgisine
yer verilmiştir. Notaların parmak ve dil uyumu ile nasıl şekilleneceği ve nasıl bir çalışma disiplini
içinde gerçekleşeceği planlanmıştır. Kamış-nota-diyafram-parmak-dil-nüans ve disiplin
kavramlarının staccato tekniğinde ayrılmaz bir bütün olduğu saptanmıştır. Araştırmada literatür
taraması yapılarak staccato ile ilgili çalışmalar taranmıştır. Tarama sonucunda klarnet tekniğin de
kullanılan staccato eser çalışmasının az olduğu tespit edilmiştir. Metot taramasında da etüt
çalışmasının daha çok olduğu saptanmıştır. Böylelikle klarnetin staccato tekniğini hızlandırma ve
güçlendirme çalışmaları sunulmuştur. Staccato etüt çalışmaları yapılırken, araya eser çalışmasının
girmesi beyni rahatlattığı ve istekliliği daha arttırdığı gözlemlenmiştir. Staccato çalışmasını yaparken
doğru bir kamış seçimi üzerinde de durulmuştur. Staccato tekniğini doğru çalışmak için doğru bir
kamışın dil hızını arttırdığı saptanmıştır. Doğru bir kamış seçimi kamıştan rahat ses çıkmasına
bağlıdır. Kamış, dil atma gücünü vermiyorsa daha doğru bir kamış seçiminin yapılması gerekliliği
vurgulanmıştır. Staccato çalışmalarında baştan sona bir eseri yorumlamak zor olabilir. Bu açıdan
çalışma, verilen müzikal nüanslara uymanın, dil atış performansını rahatlattığını ortaya koymuştur.
Gelecek nesillere edinilen bilgi ve birikimlerin aktarılması ve geliştirici olması teşvik edilmiştir.
Çıkacak eserlerin nasıl çözüleceği, staccato tekniğinin nasıl üstesinden gelinebileceği anlatılmıştır.
Staccato tekniğinin daha kısa sürede çözüme kavuşturulması amaç edinilmiştir. Parmakların
yerlerini öğrettiğimiz kadar belleğimize de çalışmaların kaydedilmesi önemlidir. Gösterilen azmin ve
sabrın sonucu olarak ortaya çıkan yapıt başarıyı daha da yukarı seviyelere çıkaracaktır
Learning disentangled speech representations
A variety of informational factors are contained within the speech signal and a single short recording of speech reveals much more than the spoken words. The best method to extract and represent informational factors from the speech signal ultimately depends on which informational factors are desired and how they will be used. In addition, sometimes methods will capture more than one informational factor at the same time such as speaker identity, spoken content, and speaker prosody.
The goal of this dissertation is to explore different ways to deconstruct the speech signal into abstract representations that can be learned and later reused in various speech technology tasks. This task of deconstructing, also known as disentanglement, is a form of distributed representation learning. As a general approach to disentanglement, there are some guiding principles that elaborate what a learned representation should contain as well as how it should function. In particular, learned representations should contain all of the requisite information in a more compact manner, be interpretable, remove nuisance factors of irrelevant information, be useful in downstream tasks, and independent of the task at hand. The learned representations should also be able to answer counter-factual questions.
In some cases, learned speech representations can be re-assembled in different ways according to the requirements of downstream applications. For example, in a voice conversion task, the speech content is retained while the speaker identity is changed. And in a content-privacy task, some targeted content may be concealed without affecting how surrounding words sound. While there is no single-best method to disentangle all types of factors, some end-to-end approaches demonstrate a promising degree of generalization to diverse speech tasks.
This thesis explores a variety of use-cases for disentangled representations including phone recognition, speaker diarization, linguistic code-switching, voice conversion, and content-based privacy masking. Speech representations can also be utilised for automatically assessing the quality and authenticity of speech, such as automatic MOS ratings or detecting deep fakes. The meaning of the term "disentanglement" is not well defined in previous work, and it has acquired several meanings depending on the domain (e.g. image vs. speech). Sometimes the term "disentanglement" is used interchangeably with the term "factorization". This thesis proposes that disentanglement of speech is distinct, and offers a viewpoint of disentanglement that can be considered both theoretically and practically
Quantitative Study of Predictive Relationships Between English Language Proficiency, Academic Growth, and Academic Achievement Assessments in North Georgia
This study examined the predictive relationship between ELs' proficiency levels on the Assessing Comprehension and Communication in English State-to-State for ELs 2.0, students' performance on English language arts Georgia Milestones Assessment System, and academic growth on the Measures of Academic Progress. It was comprised of third through fifth grade English Language Learners. The study compared the percentage of English language learner students at each proficiency level, gender, and grade level and their achievement of English language arts on the Georgia Milestones Assessment System and growth from the beginning of the year to the end of the year on Measure of Academic Progress. The study was evaluated by conducting Pearson correlation coefficients, one-way ANOVA, and mediation analysis.
Results for this research question indicated a significant positive relationship between academic achievement and academic growth. There was a significant positive relationship between academic achievement and all eight domains of English proficiency. The results indicated as grade level increased, English proficiency increased, and academic growth and achievement decreased. Results for this research question indicated a significant effect on all eight domains of English proficiency. The results indicated academic achievement is not obtained for almost 77% of ELs scoring in the 4.3 – 4.9 English proficiency level. There were significant results for all eight domains of English proficiency and academic achievement. The three domains of speaking, oral, and composite were mediated by academic growth.
Keywords: English Language Proficiency; Academic Growth; Academic Achievement Assessments; English Language Learners; ELLs; EL;ABSTRACT i -- TABLE OF CONTENTS ii -- LIST OF TABLES vi -- LIST OF FIGURES vii -- ACKNOWLEDGMENT viii -- DEDICATION ix -- Chapter I INTRODUCTION 1 -- Overview 1 -- Statement of the Problem 4 -- Purpose 5 -- Research Questions 7 -- Significance of the Study 8 -- Conceptual Framework 9 -- Summary of Methodology 13 -- Limitations 14 -- Summary 14 -- Definitions of Terms 15 -- Chapter II REVIEW OF LITERATURE 17 -- Overview 17 -- ELs 17 -- Accommodations 18 -- Exit Criteria 19 -- Post Exit Monitoring 20 -- ELs in the Classroom: Best Practices 21 -- Challenges in the Classroom 21 -- The Role of Leadership 24 -- State and Federal Mandates 25 -- Standards 26 -- Every Student Succeed Act 26 -- Georgia Standards of Excellence 27 -- World-Class Instructional Design Assessment Consortium 27 -- English Language Proficiency: Years to Proficiency 28 -- Gender 30 -- Reclassifying ELs 31 -- Language Proficiency Assessment 35 -- Academic Achievement Assessment 36 -- Adaptive Academic Assessment 36 -- Summary 37 -- Chapter III METHODOLOGY 39 -- Overview 39 -- Research Questions 40 -- Research Design 40 -- Sample 42 -- Description of the Population 43 -- Data Collection 44 -- Procedures 47 -- Threats to Validity 50 -- Summary 50 -- Chapter IV RESULTS 52 -- Data Analysis 53 -- Descriptive Statistics 54 -- Results by Questions 55 -- Research Question 1 55 -- Research Question 2 64 -- Research Question 3 69 -- Summary 79 -- Chapter V SUMMARY AND DISCUSSION 82 -- Overview 82 -- Overview of the Sample and Data Collection 86 -- Quantitative Findings 87 -- Implications of Findings 89 -- Limitations to the Study 94 -- Recommendations for Future Research 95 -- Summary 96 -- REFERENCES 98 -- Appendix A Institutional Review Board Protocol Exemption Report 109 -- Appendix B Letter of Cooperation 1 111 -- Appendix C Letter of Cooperation 2 113 -- Appendix D Data Collection 115Bochenko, MichaelSakhavat, MammadovHsiao, E-LingHill, D. LaverneEd.D.Education in Leadershi
Towards a more just refuge regime: quotas, markets and a fair share
The international refugee regime is beset by two problems: Responsibility for refuge falls
disproportionately on a few states and many owed refuge do not get it. In this work, I explore
remedies to these problems. One is a quota distribution wherein states are distributed
responsibilities via allotment. Another is a marketized quota system wherein states are free to buy
and sell their allotments with others. I explore these in three parts. In Part 1, I develop the prime
principles upon which a just regime is built and with which alternatives can be adjudicated. The
first and most important principle – ‘Justice for Refugees’ – stipulates that a just regime provides
refuge for all who have a basic interest in it. The second principle – ‘Justice for States’ – stipulates
that a just distribution of refuge responsibilities among states is one that is capacity considerate. In
Part 2, I take up several vexing questions regarding the distribution of refuge responsibilities
among states in a collective effort. First, what is a state’s ‘fair share’? The answer requires the
determination of some logic – some metric – with which a distribution is determined. I argue that
one popular method in the political theory literature – a GDP-based distribution – is normatively
unsatisfactory. In its place, I posit several alternative metrics that are more attuned with the
principles of justice but absent in the political theory literature: GDP adjusted for Purchasing
Power Parity and the Human Development Index. I offer an exploration of both these. Second,
are states required to ‘take up the slack’ left by defaulting peers? Here, I argue that duties of help
remain intact in cases of partial compliance among states in the refuge regime, but that political
concerns may require that such duties be applied with caution. I submit that a market instrument
offers one practical solution to this problem, as well as other advantages. In Part 3, I take aim at
marketization and grapple with its many pitfalls: That marketization is commodifying, that it is
corrupting, and that it offers little advantage in providing quality protection for refugees. In
addition to these, I apply a framework of moral markets developed by Debra Satz. I argue that a
refuge market may satisfy Justice Among States, but that it is violative of the refugees’ welfare
interest in remaining free of degrading and discriminatory treatment
Data-to-text generation with neural planning
In this thesis, we consider the task of data-to-text generation, which takes non-linguistic
structures as input and produces textual output. The inputs can take the form of
database tables, spreadsheets, charts, and so on. The main application of data-to-text
generation is to present information in a textual format which makes it accessible to
a layperson who may otherwise find it problematic to understand numerical figures.
The task can also automate routine document generation jobs, thus improving human
efficiency. We focus on generating long-form text, i.e., documents with multiple paragraphs. Recent approaches to data-to-text generation have adopted the very successful
encoder-decoder architecture or its variants. These models generate fluent (but often
imprecise) text and perform quite poorly at selecting appropriate content and ordering
it coherently. This thesis focuses on overcoming these issues by integrating content
planning with neural models. We hypothesize data-to-text generation will benefit from
explicit planning, which manifests itself in (a) micro planning, (b) latent entity planning, and (c) macro planning. Throughout this thesis, we assume the input to our
generator are tables (with records) in the sports domain. And the output are summaries
describing what happened in the game (e.g., who won/lost, ..., scored, etc.).
We first describe our work on integrating fine-grained or micro plans with data-to-text generation. As part of this, we generate a micro plan highlighting which records
should be mentioned and in which order, and then generate the document while taking
the micro plan into account.
We then show how data-to-text generation can benefit from higher level latent entity planning. Here, we make use of entity-specific representations which are dynam ically updated. The text is generated conditioned on entity representations and the
records corresponding to the entities by using hierarchical attention at each time step.
We then combine planning with the high level organization of entities, events, and
their interactions. Such coarse-grained macro plans are learnt from data and given
as input to the generator. Finally, we present work on making macro plans latent
while incrementally generating a document paragraph by paragraph. We infer latent
plans sequentially with a structured variational model while interleaving the steps of
planning and generation. Text is generated by conditioning on previous variational
decisions and previously generated text.
Overall our results show that planning makes data-to-text generation more interpretable, improves the factuality and coherence of the generated documents and re duces redundancy in the output document
'Exarcheia doesn't exist': Authenticity, Resistance and Archival Politics in Athens
My thesis investigates the ways people, materialities and urban spaces interact to form affective ecologies and produce historicity. It focuses on the neighbourhood of Exarcheia, Athens’ contested political topography par excellence, known for its production of radical politics of discontent and resistance to state oppression and eoliberal capitalism. Embracing Exarcheia’s controversial status within Greek vernacular, media and state discourses, this thesis aims to unpick the neighbourhoods’ socio-spatial assemblage imbued with affect and formed through the numerous (mis)understandings and (mis)interpretations rooted in its turbulent political history. Drawing on theory on urban spaces, affect, hauntology and archival politics, I argue for Exarcheia as an unwavering archival space composed of affective chronotopes – (in)tangible loci that defy space and temporality. I posit that the interwoven narratives and materialities emerging in my fieldwork are persistently – and perhaps obsessively – reiterating themselves and remaining imprinted on the neighbourhood’s landscape as an incessant reminder of violent histories that the state often seeks to erase and forget. Through this analysis, I contribute to understandings of place as a primary ethnographic ‘object’ and the ways in which place forms complex interactions and relationships with social actors, shapes their subjectivities, retains and bestows their memories and senses of historicity
A Case Study Examining Japanese University Students' Digital Literacy and Perceptions of Digital Tools for Academic English learning
Current Japanese youth are constantly connected to the Internet and using digital devices, but predominantly for social media and entertainment. According to literature on the Japanese digital native, tertiary students do not—and cannot—use technology with any reasonable fluency, but the likely reasons are rarely addressed. To fill the gap in the literature, this study, by employing a case study methodology, explores students’ experience with technology for English learning through the introduction of digital tools. First-year Japanese university students in an Academic English Program (AEP) were introduced to a variety of easily available digital tools. The instruction was administered online, and each tool was accompanied by a task directly related to classwork. Both quantitative and qualitative data were collected in the form of a pre-course Computer Literacy Survey, a post-course open-ended Reflection Activity survey, and interviews. The qualitative data was reviewed drawing on the Technology Acceptance Model (TAM) and its educational variants as an analytical framework. Educational, social, and cultural factors were also examined to help identify underlying factors that would influence students’ perceptions. The results suggest that the subjects’ lack of awareness of, and experience with, the use of technology for learning are the fundamental causes of their perceptions of initial difficulty. Based on these findings, this study proposes a possible technology integration model that enhances digital literacy for more effective language learning in the context of Japanese education
Foundations for programming and implementing effect handlers
First-class control operators provide programmers with an expressive and efficient
means for manipulating control through reification of the current control state as a first-class object, enabling programmers to implement their own computational effects and
control idioms as shareable libraries. Effect handlers provide a particularly structured
approach to programming with first-class control by naming control reifying operations
and separating from their handling.
This thesis is composed of three strands of work in which I develop operational
foundations for programming and implementing effect handlers as well as exploring
the expressive power of effect handlers.
The first strand develops a fine-grain call-by-value core calculus of a statically
typed programming language with a structural notion of effect types, as opposed to the
nominal notion of effect types that dominates the literature. With the structural approach,
effects need not be declared before use. The usual safety properties of statically typed
programming are retained by making crucial use of row polymorphism to build and
track effect signatures. The calculus features three forms of handlers: deep, shallow,
and parameterised. They each offer a different approach to manipulate the control state
of programs. Traditional deep handlers are defined by folds over computation trees,
and are the original con-struct proposed by Plotkin and Pretnar. Shallow handlers are
defined by case splits (rather than folds) over computation trees. Parameterised handlers
are deep handlers extended with a state value that is threaded through the folds over
computation trees. To demonstrate the usefulness of effects and handlers as a practical
programming abstraction I implement the essence of a small UNIX-style operating
system complete with multi-user environment, time-sharing, and file I/O.
The second strand studies continuation passing style (CPS) and abstract machine
semantics, which are foundational techniques that admit a unified basis for implementing deep, shallow, and parameterised effect handlers in the same environment. The
CPS translation is obtained through a series of refinements of a basic first-order CPS
translation for a fine-grain call-by-value language into an untyped language. Each refinement moves toward a more intensional representation of continuations eventually
arriving at the notion of generalised continuation, which admit simultaneous support for
deep, shallow, and parameterised handlers. The initial refinement adds support for deep
handlers by representing stacks of continuations and handlers as a curried sequence of
arguments. The image of the resulting translation is not properly tail-recursive, meaning some function application terms do not appear in tail position. To rectify this the
CPS translation is refined once more to obtain an uncurried representation of stacks
of continuations and handlers. Finally, the translation is made higher-order in order to
contract administrative redexes at translation time. The generalised continuation representation is used to construct an abstract machine that provide simultaneous support for
deep, shallow, and parameterised effect handlers. kinds of effect handlers.
The third strand explores the expressiveness of effect handlers. First, I show that
deep, shallow, and parameterised notions of handlers are interdefinable by way of typed
macro-expressiveness, which provides a syntactic notion of expressiveness that affirms
the existence of encodings between handlers, but it provides no information about the
computational content of the encodings. Second, using the semantic notion of expressiveness I show that for a class of programs a programming language with first-class
control (e.g. effect handlers) admits asymptotically faster implementations than possible in a language without first-class control
- …