2,290 research outputs found

    Achieving an appropriate balance between precision, support, and comprehensibility in the evolution of classification rules

    Get PDF
    This article proposes a method for achieving an appropriate balance between the parameters of support, precision, and simplicity during the evolution of classification rules by means of genetic programming. The method includes an adaptive procedure in order to achieve such balance. This work lies within the data mining context, more precisely, it focuses on the extraction of comprehensible knowledge where the approach introduced plays a predominant role. Experimental results demonstrate the advantages of using the proposed methodRed de Universidades con Carreras en Informática (RedUNCI

    Achieving an appropriate balance between precision, support, and comprehensibility in the evolution of classification rules

    Get PDF
    This article proposes a method for achieving an appropriate balance between the parameters of support, precision, and simplicity during the evolution of classification rules by means of genetic programming. The method includes an adaptive procedure in order to achieve such balance. This work lies within the data mining context, more precisely, it focuses on the extraction of comprehensible knowledge where the approach introduced plays a predominant role. Experimental results demonstrate the advantages of using the proposed methodRed de Universidades con Carreras en Informática (RedUNCI

    A Performance-Explainability-Fairness Framework For Benchmarking ML Models

    Get PDF
    Machine learning (ML) models have achieved remarkable success in various applications; however, ensuring their robustness and fairness remains a critical challenge. In this research, we present a comprehensive framework designed to evaluate and benchmark ML models through the lenses of performance, explainability, and fairness. This framework addresses the increasing need for a holistic assessment of ML models, considering not only their predictive power but also their interpretability and equitable deployment. The proposed framework leverages a multi-faceted evaluation approach, integrating performance metrics with explainability and fairness assessments. Performance evaluation incorporates standard measures such as accuracy, precision, and recall, but extends to overall balanced error rate, overall area under the receiver operating characteristic (ROC) curve (AUC), to capture model behavior across different performance aspects. Explainability assessment employs state-of-the-art techniques to quantify the interpretability of model decisions, ensuring that model behavior can be understood and trusted by stakeholders. The fairness evaluation examines model predictions in terms of demographic parity, equalized odds, thereby addressing concerns of bias and discrimination in the deployment of ML systems. To demonstrate the practical utility of the framework, we apply it to a diverse set of ML algorithms across various functional domains, including finance, criminology, education, and healthcare prediction. The results showcase the importance of a balanced evaluation approach, revealing trade-offs between performance, explainability, and fairness that can inform model selection and deployment decisions. Furthermore, we provide insights into the analysis of tradeoffs in selecting the appropriate model for use cases where performance, interpretability and fairness are important. In summary, the Performance-Explainability-Fairness Framework offers a unified methodology for evaluating and benchmarking ML models, enabling practitioners and researchers to make informed decisions about model suitability and ensuring responsible and equitable AI deployment. We believe that this framework represents a crucial step towards building trustworthy and accountable ML systems in an era where AI plays an increasingly prominent role in decision-making processes

    Instructors’ Views towards the Second Language Acquisition of the Spanish Subjunctive

    Get PDF
    abstract: The study of Spanish instructors’ beliefs is a recent development and the body of work is small with little research conducted on their insights on the acquisition of any grammar form. Still, Spanish grammar includes the notoriously difficult subjunctive, a grammatical irrealis mood that is affixed to verbs. A national survey was conducted on Spanish professors and instructors (N=73) who teach at institutions randomly selected from a representative sample of American institutions of higher education. The survey was conducted to inquire on their beliefs regarding the most complex forms in Spanish, the causes of the subjunctive difficulty, and their preferred methods of teaching the form. The results first indicate that participants rated the subjunctive the most difficult grammar form. They attributed the cause of difficulty to be primarily interference from the first language and its abstractness. For instructing the subjunctive, participants generally supported form-oriented instruction with a metalanguage approach that focuses on forms. However, the participants disagreed greatly on whether meaning-focused instruction was valuable and dismissed drilling instruction of the subjunctive. Data from the participants provides a distribution of overextended tense, moods, and aspects in lieu of the Spanish subjunctive. However, instructors indicated that their students’ competence of the subjunctive was higher than their performance and that comprehension was not necessarily reliant on correct usage of the subjunctive as it was for proficiency. Moreover, they provided qualitative data of effective methods and pedagogical challenges of the subjunctive. This study illuminates some of the contributing factors of subjunctive difficulty and preferred pedagogical approaches for teaching it. It also has implications that meaning may not be obstructed if students do not use subjunctive.Dissertation/ThesisMasters Thesis Linguistics and Applied Linguistics 201

    Mathematically aggregating experts' predictions of possible futures

    Get PDF
    Structured protocols offer a transparent and systematic way to elicit and combine/aggregate, probabilistic predictions from multiple experts. These judgements can be aggregated behaviourally or mathematically to derive a final group prediction. Mathematical rules (e.g., weighted linear combinations of judgments) provide an objective approach to aggregation. The quality of this aggregation can be defined in terms of accuracy, calibration and informativeness. These measures can be used to compare different aggregation approaches and help decide on which aggregation produces the “best” final prediction. When experts’ performance can be scored on similar questions ahead of time, these scores can be translated into performance-based weights, and a performance-based weighted aggregation can then be used. When this is not possible though, several other aggregation methods, informed by measurable proxies for good performance, can be formulated and compared. Here, we develop a suite of aggregation methods, informed by previous experience and the available literature. We differentially weight our experts’ estimates by measures of reasoning, engagement, openness to changing their mind, informativeness, prior knowledge, and extremity, asymmetry or granularity of estimates. Next, we investigate the relative performance of these aggregation methods using three datasets. The main goal of this research is to explore how measures of knowledge and behaviour of individuals can be leveraged to produce a better performing combined group judgment. Although the accuracy, calibration, and informativeness of the majority of methods are very similar, a couple of the aggregation methods consistently distinguish themselves as among the best or worst. Moreover, the majority of methods outperform the usual benchmarks provided by the simple average or the median of estimates

    Consolidation of Customized Product Copies into Software Product Lines

    Get PDF
    In software development, project constraints lead to customer-specific variants by copying and adapting the product. During this process, modifications are scattered all over the code. Although this is flexible and efficient in the short term, a Software Product Line (SPL) offers better results in the long term, regarding cost reduction, time-to-market, and quality attributes. This book presents a novel approach named SPLevo, which consolidates customized product copies into an SPL

    Evolutionary Learning of Fuzzy Rules for Regression

    Get PDF
    The objective of this PhD Thesis is to design Genetic Fuzzy Systems (GFS) that learn Fuzzy Rule Based Systems to solve regression problems in a general manner. Particularly, the aim is to obtain models with low complexity while maintaining high precision without using expert-knowledge about the problem to be solved. This means that the GFSs have to work with raw data, that is, without any preprocessing that help the learning process to solve a particular problem. This is of particular interest, when no knowledge about the input data is available or for a first approximation to the problem. Moreover, within this objective, GFSs have to cope with large scale problems, thus the algorithms have to scale with the data

    Consolidation of Customized Product Copies into Software Product Lines

    Get PDF
    In software development, project constraints lead to customer-specific variants by copying and adapting the product. During this process, modifications are scattered all over the code. Although this is flexible and efficient in the short term, a Software Product Line (SPL) offers better results in the long term, regarding cost reduction, time-to-market, and quality attributes. This book presents a novel approach named SPLevo, which consolidates customized product copies into an SPL

    Dimensionality Reduction of Quality Objectives for Web Services Design Modularization

    Full text link
    With the increasing use of service-oriented Architecture (SOA) in new software development, there is a growing and urgent need to improve current practice in service-oriented design. To improve the design of Web services, the search for Web services interface modularization solutions deals, in general, with a large set of conflicting quality metrics. Deciding about which and how the quality metrics are used to evaluate generated solutions are always left to the designer. Some of these objectives could be correlated or conflicting. In this paper, we propose a dimensionality reduction approach based on Non-dominated Sorting Genetic Algorithm (NSGA-II) to address the Web services re-modularization problem. Our approach aims at finding the best-reduced set of objectives (e.g. quality metrics) that can generate near optimal Web services modularization solutions to fix quality issues in Web services interface. The algorithm starts with a large number of interface design quality metrics as objectives (e.g. coupling, cohesion, number of ports, number of port types, and number of antipatterns) that are reduced based on the nonlinear correlation information entropy (NCIE).The statistical analysis of our results, based on a set of 22 real world Web services provided by Amazon and Yahoo, confirms that our dimensionality reduction Web services interface modularization approach reduced significantly the number of objectives on several case studies to a minimum of 2 objectives and performed significantly better than the state-of-the-art modularization techniques in terms of generating well-designed Web services interface for users.Master of ScienceSoftware Engineering, College of Engineering & Computer ScienceUniversity of Michigan-Dearbornhttps://deepblue.lib.umich.edu/bitstream/2027.42/145687/1/Thesis Report_Hussein Skaf.pdfDescription of Thesis Report_Hussein Skaf.pdf : Thesi
    corecore