3,887 research outputs found

    An Architecture of an Intelligent Tutoring System to Support Distance Learning

    Get PDF
    This paper outlines a design framework of an intelligent tutoring system (ITS). ITS focuses on a newer and more comprehensive distance learning (DL) process as compared to the established traditional DL programs practiced today. The DL model presented in this paper (CHARLIE) is a high level software based tutorial that has the ability to encompass a wide variety of current DL technologies in a single DL session. CHARLIE's architecture has four components: Control Component (responsible for the interaction between software agents and the operating system); Instructional Component (concerned with the instructional aspects of an ITS session); Text Analysis Component (analyzes the partial syntax and partial semantics of the text in the session); Student Modeling Component (analyzes a student's progress and determines the best model for learning during a session). Each component is serviced by a set of software agents to accomplish its mission. Three additional entities in CHARLIE are two separate databases and an explanation facility. Six agents have been implemented in CHARLIE to create a DL course in C++ programming. Much of CHARLIE remains to be completed which opens many areas for research

    Proceedings of the 1993 Conference on Intelligent Computer-Aided Training and Virtual Environment Technology, Volume 1

    Get PDF
    These proceedings are organized in the same manner as the conference's contributed sessions, with the papers grouped by topic area. These areas are as follows: VE (virtual environment) training for Space Flight, Virtual Environment Hardware, Knowledge Aquisition for ICAT (Intelligent Computer-Aided Training) & VE, Multimedia in ICAT Systems, VE in Training & Education (1 & 2), Virtual Environment Software (1 & 2), Models in ICAT systems, ICAT Commercial Applications, ICAT Architectures & Authoring Systems, ICAT Education & Medical Applications, Assessing VE for Training, VE & Human Systems (1 & 2), ICAT Theory & Natural Language, ICAT Applications in the Military, VE Applications in Engineering, Knowledge Acquisition for ICAT, and ICAT Applications in Aerospace

    ICLabel: An automated electroencephalographic independent component classifier, dataset, and website

    Full text link
    The electroencephalogram (EEG) provides a non-invasive, minimally restrictive, and relatively low cost measure of mesoscale brain dynamics with high temporal resolution. Although signals recorded in parallel by multiple, near-adjacent EEG scalp electrode channels are highly-correlated and combine signals from many different sources, biological and non-biological, independent component analysis (ICA) has been shown to isolate the various source generator processes underlying those recordings. Independent components (IC) found by ICA decomposition can be manually inspected, selected, and interpreted, but doing so requires both time and practice as ICs have no particular order or intrinsic interpretations and therefore require further study of their properties. Alternatively, sufficiently-accurate automated IC classifiers can be used to classify ICs into broad source categories, speeding the analysis of EEG studies with many subjects and enabling the use of ICA decomposition in near-real-time applications. While many such classifiers have been proposed recently, this work presents the ICLabel project comprised of (1) an IC dataset containing spatiotemporal measures for over 200,000 ICs from more than 6,000 EEG recordings, (2) a website for collecting crowdsourced IC labels and educating EEG researchers and practitioners about IC interpretation, and (3) the automated ICLabel classifier. The classifier improves upon existing methods in two ways: by improving the accuracy of the computed label estimates and by enhancing its computational efficiency. The ICLabel classifier outperforms or performs comparably to the previous best publicly available method for all measured IC categories while computing those labels ten times faster than that classifier as shown in a rigorous comparison against all other publicly available EEG IC classifiers.Comment: Intended for NeuroImage. Updated from version one with minor editorial and figure change

    THE APPLICATION OF COMPUTER VISION, MACHINE AND DEEP LEARNING ALGORITHMS UTILIZING MATLAB

    Get PDF
    MATLAB is a multi-paradigm proprietary programming language and numerical computing environment developed by MathWorks. Within MATLAB Integrated Development Environment (IDE) you can perform Computer-aided design (CAD), different matrix manipulations, plotting of functions and data, implementation algorithms, creation of user interfaces, and has the ability to interface with programs written in other languages1. Since, its launch in 1984 MATLAB software has not particularly been associated within the field of data science. In 2013, that changed with the launch of their new data science concentrated toolboxes that included Deep Learning, Image Processing, Computer Vision, and then a year later Statistics and Machine Learning. The main objective of my thesis was to research and explore the field of data science. More specifically pertaining to the development of an object recognition application that could be built entirely using MATLAB IDE and have a positive social impact on the deaf community. And in doing so, answering the question, could MATLAB be utilized for development of this type of application? To simultaneously answer this question while addressing my main objectives, I constructed two different object recognition protocols utilizing MATLAB_R2019 with the add-on data science tool packages. I named the protocols ASLtranslate (I) and (II). This allowed me to experiment with all of MATLAB data science toolboxes while learning the differences, benefits, and disadvantages of using multiple approaches to the same problem. The methods and approaches for the design of both versions was very similar. ASLtranslate takes in 2D image of American Sign Language (ASL) hand gestures as an input, classifies the image and then outputs its corresponding alphabet character. ASLtranslate (I) was an implementation of image category classification using machine learning methods. ASLtranslate (II) was implemented by using a deep learning method called transfer learning, done by fine-tuning a pre-trained convolutional neural network (CNN), AlexNet, to perform classification on a new collection of images

    Assessing Adaptive Learning Styles in Computer Science Through a Virtual World

    Get PDF
    abstract: Programming is quickly becoming as ubiquitous and essential a skill as general mathematics. However, many elementary and high school students are still not aware of what the computer science field entails. To make matters worse, students who are introduced to computer science are frequently being fed only part of what it is about rather than its entire construction. Consequently, they feel out of their depth when they approach college. Research has discovered that by teaching computer science and programming through a problem-driven approach and focusing on a combination of syntax and computational thinking, students can be prepared when entering higher levels of computer science education. This thesis describes the design, development, and early user testing of a theory-based virtual world for computer science instruction called System Dot. System Dot was designed to visually manifest programming instructions into interactable objects, giving players a way to see coding as tangible entities rather than text on a white screen. In order for System Dot to convey the true nature of computer science, a custom predictive recursive descent parser was embedded in the program to validate any user-generated solutions to pre-defined logical platforming puzzles. Steps were taken to adapt the virtual world to player behavior by creating a system to detect their learning style playing the game. Through a dynamic Bayesian network, System Dot aims to classify a player’s learning style based on the Felder-Sylverman Learning Style Model (FSLSM). Testers played through the first half of System Dot, which was enough to test out the Bayesian network and initial learning style classification. This classification was then compared to the assessment by Felder’s Index of Learning Styles Questionnaire (ILSQ). Lastly, this thesis will also discuss ways to use the results from the user testing to implement a personalized feedback system for the virtual world in the future and what has been learned through the learning style method.Dissertation/ThesisMasters Thesis Computer Science 201

    Tutoring Students with Adaptive Strategies

    Get PDF
    Adaptive learning is a crucial part in intelligent tutoring systems. It provides students with appropriate tutoring interventions, based on students’ characteristics, status, and other related features, in order to optimize their learning outcomes. It is required to determine students’ knowledge level or learning progress, based on which it then uses proper techniques to choose the optimal interventions. In this dissertation work, I focus on these aspects related to the process in adaptive learning: student modeling, k-armed bandits, and contextual bandits. Student modeling. The main objective of student modeling is to develop cognitive models of students, including modeling content skills and knowledge about learning. In this work, we investigate the effect of prerequisite skill in predicting students’ knowledge in post skills, and we make use of the prerequisite performance in different student models. As a result, this makes them superior to traditional models. K-armed bandits. We apply k-armed bandit algorithms to personalize interventions for students, to optimize their learning outcomes. Due to the lack of diverse interventions and small difference of intervention effectiveness in educational experiments, we also propose a simple selection strategy, and compare it with several k-armed bandit algorithms. Contextual bandits. In contextual bandit problem, additional side information, also called context, can be used to determine which action to select. First, we construct a feature evaluation mechanism, which determines which feature to be combined with bandits. Second, we propose a new decision tree algorithm, which is capable of detecting aptitude treatment effect for students. Third, with combined bandits with the decision tree, we apply the contextual bandits to make personalization in two different types of data, simulated data and real experimental data

    Explainable Hopfield Neural Networks Using an Automatic Video-Generation System

    Get PDF
    Hopfield Neural Networks (HNNs) are recurrent neural networks used to implement associative memory. They can be applied to pattern recognition, optimization, or image segmentation. However, sometimes it is not easy to provide the users with good explanations about the results obtained with them due to mainly the large number of changes in the state of neurons (and their weights) produced during a problem of machine learning. There are currently limited techniques to visualize, verbalize, or abstract HNNs. This paper outlines how we can construct automatic video-generation systems to explain its execution. This work constitutes a novel approach to obtain explainable artificial intelligence systems in general and HNNs in particular building on the theory of data-to-text systems and software visualization approaches. We present a complete methodology to build these kinds of systems. Software architecture is also designed, implemented, and tested. Technical details about the implementation are also detailed and explained. We apply our approach to creating a complete explainer video about the execution of HNNs on a small recognition problem. Finally, several aspects of the videos generated are evaluated (quality, content, motivation and design/presentation).University of the Bio-Bio. Vicerrectoria de Investigacion. Facultad de Ciencias Empresariales. Departamento de Sistemas de Informacion

    Insight in problem solving : developing a neural network theoretical account of the processes involved in attaining insight

    Get PDF
    Bibliography: leaves 151-164.Insight has enjoyed the reputation of an elusive phenomenon in psychology and insight problems are very difficult to solve. Only very specific hints concerning their solution have been found to significantly increase the number of problem solvers who are able to solve insight problems. The result of this has been to suggest that insight does not exist, that it is a mysterious phenomenon, or that it is an aspect of problem solving which we have so far failed to understand. Insight in problem solving is investigated from the perspective that the phenomenon needs explanation and it is argued that, while insight has been operationally defined and a clear set of key empirical findings have been established, the conceptual explanation of insight has been largely ignored. It is suggested that a conceptual account of insight is needed so that this aspect of cognitive processing can be incorporated into the main body of cognitive research on problem solving. The current tension in cognitive science and cognitive psychology is examined and it is argued that writing a conceptual account of insight in neural network theoretical terms will not only advance our understanding of insight, but will also reflect on the debate in cognitive theory. This is a result of its status as an aspect of problem solving and as a phenomenon which symbolic theory has so far failed to offer a clear explanation for. A conceptual account of insight in neural network terms is advanced which offers a comprehensive account of the key empirical findings on insight. It is suggested that insight can be understood as the recognition of a pattern to insight problems. Predictions derived from the theory suggest that overcoming the effects of past learning, employing conceptual transfer, and fostering expertise at insight problem solving will significantly facilitate insightful problem solution

    iFocus: A Framework for Non-intrusive Assessment of Student Attention Level in Classrooms

    Get PDF
    The process of learning is not merely determined by what the instructor teaches, but also by how the student receives that information. An attentive student will naturally be more open to obtaining knowledge than a bored or frustrated student. In recent years, tools such as skin temperature measurements and body posture calculations have been developed for the purpose of determining a student\u27s affect, or emotional state of mind. However, measuring eye-gaze data is particularly noteworthy in that it can collect measurements non-intrusively, while also being relatively simple to set up and use. This paper details how data obtained from such an eye-tracker can be used to predict a student\u27s attention as a measure of affect over the course of a class. From this research, an accuracy of 77% was achieved using the Extreme Gradient Boosting technique of machine learning. The outcome indicates that eye-gaze can be indeed used as a basis for constructing a predictive model
    • …
    corecore