43 research outputs found
Extension des modèles de prédiction de la qualité du logiciel en utilisant la logique floue et les heuristiques du domaine
Mémoire numérisé par la Direction des bibliothèques de l'Université de Montréal
Facial image pre-processing and emotion classification: A deep learning approach
© 2019 IEEE. Facial emotion detection and expressions are vital for applications that require credibility assessment, evaluating truthfulness, and detection of deception. However, most of the research reveal low accuracy in emotion detection mainly due to the low quality of images under consideration. Conducting intensive pre-processing activities and using artificial intelligence especially deep learning techniques are increasing accuracy in computational predictions. Our research focuses on emotion detection using deep learning techniques and combined preprocessing activities. We propose a solution that applies and compares four deep learning models for image pre-processing with the main objective to improve emotion recognition accuracy. Our methodology includes three major stages in the data value chain, pre-processing, deep learning and post-processing. We evaluate the proposed scheme on a real facial data set, namely Facial Image Data of Indian Film Stars for our study. The experimentation compares the performance of various deep learning techniques on the facial image data and confirms that our approach enhanced significantly the image quality using intensive pre-processing and deep-learning, improves accuracy in emotion prediction
Big data quality framework: a holistic approach to continuous quality management
Big Data is an essential research area for governments, institutions, and private agencies to support their analytics decisions. Big Data refers to all about data, how it is collected, processed, and analyzed to generate value-added data-driven insights and decisions. Degradation in Data Quality may result in unpredictable consequences. In this case, confidence and worthiness in the data and its source are lost. In the Big Data context, data characteristics, such as volume, multi-heterogeneous data sources, and fast data generation, increase the risk of quality degradation and require efficient mechanisms to check data worthiness. However, ensuring Big Data Quality (BDQ) is a very costly and time-consuming process, since excessive computing resources are required. Maintaining Quality through the Big Data lifecycle requires quality profiling and verification before its processing decision. A BDQ Management Framework for enhancing the pre-processing activities while strengthening data control is proposed. The proposed framework uses a new concept called Big Data Quality Profile. This concept captures quality outline, requirements, attributes, dimensions, scores, and rules. Using Big Data profiling and sampling components of the framework, a faster and efficient data quality estimation is initiated before and after an intermediate pre-processing phase. The exploratory profiling component of the framework plays an initial role in quality profiling; it uses a set of predefined quality metrics to evaluate important data quality dimensions. It generates quality rules by applying various pre-processing activities and their related functions. These rules mainly aim at the Data Quality Profile and result in quality scores for the selected quality attributes. The framework implementation and dataflow management across various quality management processes have been discussed, further some ongoing work on framework evaluation and deployment to support quality evaluation decisions conclude the paper
Empowering Patient Similarity Networks through Innovative Data-Quality-Aware Federated Profiling
Continuous monitoring of patients involves collecting and analyzing sensory data from a multitude of sources. To overcome communication overhead, ensure data privacy and security, reduce data loss, and maintain efficient resource usage, the processing and analytics are moved close to where the data are located (e.g., the edge). However, data quality (DQ) can be degraded because of imprecise or malfunctioning sensors, dynamic changes in the environment, transmission failures, or delays. Therefore, it is crucial to keep an eye on data quality and spot problems as quickly as possible, so that they do not mislead clinical judgments and lead to the wrong course of action. In this article, a novel approach called federated data quality profiling (FDQP) is proposed to assess the quality of the data at the edge. FDQP is inspired by federated learning (FL) and serves as a condensed document or a guide for node data quality assurance. The FDQP formal model is developed to capture the quality dimensions specified in the data quality profile (DQP). The proposed approach uses federated feature selection to improve classifier precision and rank features based on criteria such as feature value, outlier percentage, and missing data percentage. Extensive experimentation using a fetal dataset split into different edge nodes and a set of scenarios were carefully chosen to evaluate the proposed FDQP model. The results of the experiments demonstrated that the proposed FDQP approach positively improved the DQ, and thus, impacted the accuracy of the federated patient similarity network (FPSN)-based machine learning models. The proposed data-quality-aware federated PSN architecture leveraging FDQP model with data collected from edge nodes can effectively improve the data quality and accuracy of the federated patient similarity network (FPSN)-based machine learning models. Our profiling algorithm used lightweight profile exchange instead of full data processing at the edge, which resulted in optimal data quality achievement, thus improving efficiency. Overall, FDQP is an effective method for assessing data quality in the edge computing environment, and we believe that the proposed approach can be applied to other scenarios beyond patient monitoring
Trustworthy Federated Learning: A Survey
Federated Learning (FL) has emerged as a significant advancement in the field
of Artificial Intelligence (AI), enabling collaborative model training across
distributed devices while maintaining data privacy. As the importance of FL
increases, addressing trustworthiness issues in its various aspects becomes
crucial. In this survey, we provide an extensive overview of the current state
of Trustworthy FL, exploring existing solutions and well-defined pillars
relevant to Trustworthy . Despite the growth in literature on trustworthy
centralized Machine Learning (ML)/Deep Learning (DL), further efforts are
necessary to identify trustworthiness pillars and evaluation metrics specific
to FL models, as well as to develop solutions for computing trustworthiness
levels. We propose a taxonomy that encompasses three main pillars:
Interpretability, Fairness, and Security & Privacy. Each pillar represents a
dimension of trust, further broken down into different notions. Our survey
covers trustworthiness challenges at every level in FL settings. We present a
comprehensive architecture of Trustworthy FL, addressing the fundamental
principles underlying the concept, and offer an in-depth analysis of trust
assessment mechanisms. In conclusion, we identify key research challenges
related to every aspect of Trustworthy FL and suggest future research
directions. This comprehensive survey serves as a valuable resource for
researchers and practitioners working on the development and implementation of
Trustworthy FL systems, contributing to a more secure and reliable AI
landscape.Comment: 45 Pages, 8 Figures, 9 Table
Novel cloud and SOA-based framework for E-health monitoring using wireless biosensors
Various and independent studies are showing that an exponential increase of chronic diseases (CDs) is exhausting governmental and private healthcare systems to an extent that some countries allocate half of their budget to healthcare systems. To benefit from the IT development, e-health monitoring and prevention approaches revealed to be among top promising solutions. In fact, well-implemented monitoring and prevention schemes have reported a decent reduction of CDs risk and have narrowed their effects, on both patients\u27 health conditions and on government budget spent on healthcare. In this paper, we propose a framework to collect patients\u27 data in real time, perform appropriate nonintrusive monitoring, and propose medical and/or life style engagements, whenever needed and appropriate. The framework, which relies on service-oriented architecture (SOA) and the Cloud, allows a seamless integration of different technologies, applications, and services. It also integrates mobile technologies to smoothly collect and communicate vital data from a patient\u27s wearable biosensors while considering the mobile devices\u27 limited capabilities and power drainage in addition to intermittent network disconnections. Then, data are stored in the Cloud and made available via SOA to allow easy access by physicians, paramedics, or any other authorized entity. A case study has been developed to evaluate the usability of the framework, and the preliminary results that have been analyzed are showing very promising results
Enforcing quality of service within web services communities
Web services are considered as an attracting distributed approach of application/services integration over the Internet. As the number of Web Services is exponentially growing and expected to do so for the next decade, the need for categorizing and/or classifying Web Services is very crucial for their success and the success of the underlying Service Oriented architecture (SOA). Categorization aims at systematizing Web Services according to their functionalities and their Quality of Service attributes. Communities of Web Services have been used to gather Web Services based on their functionalities. In fact, Web Services in a community can offer similar and/or complementary services. In this paper, we expand Web Services communities\u27 classification by adding a new support layer for Quality of Service classification. This is done through Quality of Services specification, monitoring, and adaptation of Web Services within communities. A Web Service might be admitted to a community thanks to its high Quality of Service or might be ejected from a community due to its low Quality of Service. The focus of this paper is on the design and use of a managerial community to monitor and adapt Quality of Web Services (QoWS) of managerial Web Services for other communities, Web Services providers, and Web Services clients
A cooperative approach for QoS-aware web services\u27 selection
As Web services are becoming omnipresent on the Web, quality of Web services (QoWS) support and management is becoming a hot research topic. Several frameworks for Web service selection have been proposed to support clients in selecting suitable Web services. They are very often based on a middle-tier component to make the selection decision. These solutions suffer very often from scalability problems. To deal with this issue, we propose in this paper a new architecture for Web service selection based on a federation of cooperative brokers. Each broker of the federation manages the Web services within its domain and cooperates with its peers by exchanging information about Web services, such as reputation and QoS, to better serve client requests. We have developed a prototype of the architecture and we have conducted experiments using three broker selection policies mainly “random”, “round-robin”, and “cooperative brokers” to evaluate the degree of fulfillment of clients’ requests. Preliminary results show that the best results are obtained by using the cooperative brokers’ policy
Towards an adaptive QoS-driven monitoring of cloud SaaS
The abundance of services offered by cloud providers raises the need for Quality of Service (QoS) monitoring to enforce Service Level Agreements (SLAs). In this paper, we propose a cloud-monitoring scheme, which offers flexible and dynamically reconfigurable QoS monitoring services to adapt to various cloud based service characteristics. We propose to grant cloud service consumers the option to decide whether to switch to another service provider, if the current provider frequently violates the corresponding SLA contract. For cloud service providers, the monitoring scheme provides dashboard-like indicators to continuously manage the performance of their SaaS platform and visualise the monitoring data. Our experimental evaluation involves a real cloud platform to illustrate the capability of our monitoring scheme in detecting and reporting violations of SLAs for both single and composite SaaS services. A particular emphasis on visualisation options is highlighted when revealing and reconfiguring monitoring data in a user-friendly display
Multi-tier framework for management of web services\u27 quality
Web Services are new breed of applications that endorsed large support from main vendors from industry as well as academia. As the Web Services paradigm becomes more mature, its management is crucial to its adoption and success. Existing approaches are often limited to the platforms under which management features are provided. In this chapter, we propose an approach to provide a unique central console for management of both functional and nonfunctional aspects of Web Services. In fact, we aim at the development of a framework to provide management features to providers and clients by supporting management activities all along the lifecycle. The framework allows/forces providers to consider management activities while developing their Web Services. It allows clients to select appropriate Web Services using different criteria (name, quality). Clients make also use of the framework to check if the Web Services, they are actually using or planning to use, are behaving correctly. We evaluate the Web Services management features of our framework using a composite Web Service