24 research outputs found
Routing in MobileWireless Sensor Networks: A Leader-Based Approach
This paper presents a leader-based approach to routing in Mobile Wireless Sensor Networks (MWSN). Using local information from neighbour nodes, a leader election mechanism maintains a spanning tree in order to provide the necessary adaptations for efficient routing upon the connectivity changes resulting from the mobility of sensors or sink nodes. We present two protocols following the leader election approach, which have been implemented using Castalia and OMNeT++. The protocols have been evaluated, besides other reference MWSN routing protocols, to analyse the impact of network size and node velocity on performance, which has demonstrated the validity of our approach.Research supported by the Spanish Research Council (MINECO), Grant TIN2016-79897-P, and the Department of Education, Universities and Research of the Basque Government, Grant IT980-16
Geographical area network-structural health monitoring utility computing model
In view of intensified disasters and fatalities caused by natural phenomena and geographical expansion, there is a pressing need for a more effective environment logging for a better management and urban planning. This paper proposes a novel utility computing model (UCM) for structural health monitoring (SHM) that would enable dynamic planning of monitoring systems in an efficient and cost-effective manner in form of a SHM geo-informatics system. The proposed UCM consists of networked SHM systems that send geometrical SHM variables to SHM-UCM gateways. Every gateway is routing the data to SHM-UCM servers running a geo-spatial patch health assessment and prediction algorithm. The inputs of the prediction algorithm are geometrical variables, environmental variables, and payloads. The proposed SHM-UCM is unique in terms of its capability to manage heterogeneous SHM resources. This has been tested in a case study on Qatar University (QU) in Doha Qatar, where it looked at where SHM nodes are distributed along with occupancy density in each building. This information was taken from QU routers and zone calculation models and were then compared to ideal SHM system data. Results show the effectiveness of the proposed model in logging and dynamically planning SHM.This publication was made possible by NPRP grant # 8-1781-2-725 from the Qatar National Research Fund (a member of Qatar Foundation). The publication of this article was funded by the Qatar National Library
An adaptive trust based service quality monitoring mechanism for cloud computing
Cloud computing is the newest paradigm in distributed computing that delivers computing resources over the Internet as services. Due to the attractiveness of cloud computing, the market is currently flooded with many service providers. This
has necessitated the customers to identify the right one meeting their requirements in terms of service quality. The existing monitoring of service quality has been limited only to quantification in cloud computing. On the other hand, the continuous
improvement and distribution of service quality scores have been implemented in other distributed computing paradigms but not specifically for cloud computing. This research investigates the methods and proposes mechanisms for quantifying and
ranking the service quality of service providers. The solution proposed in this thesis consists of three mechanisms, namely service quality modeling mechanism, adaptive trust computing mechanism and trust distribution mechanism for cloud computing.
The Design Research Methodology (DRM) has been modified by adding phases, means and methods, and probable outcomes. This modified DRM is used throughout this study. The mechanisms were developed and tested gradually until the expected
outcome has been achieved. A comprehensive set of experiments were carried out in a simulated environment to validate their effectiveness. The evaluation has been carried out by comparing their performance against the combined trust model and
QoS trust model for cloud computing along with the adapted fuzzy theory based trust computing mechanism and super-agent based trust distribution mechanism, which were developed for other distributed systems. The results show that the mechanisms are faster and more stable than the existing solutions in terms of reaching the final trust scores on all three parameters tested. The results presented in this thesis are significant
in terms of making cloud computing acceptable to users in verifying the performance of the service providers before making the selection
Inference-based statistical network analysis uncovers star-like brain functional architectures for internalizing psychopathology in children
To improve the statistical power for imaging biomarker detection, we propose
a latent variable-based statistical network analysis (LatentSNA) that combines
brain functional connectivity with internalizing psychopathology, implementing
network science in a generative statistical process to preserve the
neurologically meaningful network topology in the adolescents and children
population. The developed inference-focused generative Bayesian framework (1)
addresses the lack of power and inflated Type II errors in current analytic
approaches when detecting imaging biomarkers, (2) allows unbiased estimation of
biomarkers' influence on behavior variants, (3) quantifies the uncertainty and
evaluates the likelihood of the estimated biomarker effects against chance and
(4) ultimately improves brain-behavior prediction in novel samples and the
clinical utilities of neuroimaging findings. We collectively model multi-state
functional networks with multivariate internalizing profiles for 5,000 to 7,000
children in the Adolescent Brain Cognitive Development (ABCD) study with
sufficiently accurate prediction of both children internalizing traits and
functional connectivity, and substantially improved our ability to explain the
individual internalizing differences compared with current approaches. We
successfully uncover large, coherent star-like brain functional architectures
associated with children's internalizing psychopathology across multiple
functional systems and establish them as unique fingerprints for childhood
internalization
Characterization and Modelling of Composites
Composites have increasingly been used in various structural components in the aerospace, marine, automotive, and wind energy sectors. The material characterization of composites is a vital part of the product development and production process. Physical, mechanical, and chemical characterization helps developers to further their understanding of products and materials, thus ensuring quality control. Achieving an in-depth understanding and consequent improvement of the general performance of these materials, however, still requires complex material modeling and simulation tools, which are often multiscale and encompass multiphysics. This Special Issue aims to solicit papers concerning promising, recent developments in composite modeling, simulation, and characterization, in both design and manufacturing areas, including experimental as well as industrial-scale case studies. All submitted manuscripts will undergo a rigorous review process and will only be considered for publication if they meet journal standards. Selected top articles may have their processing charges waived at the recommendation of reviewers and the Guest Editor
Un système multi-agents pour la gestion des connaissances hétérogènes et distribuées
La gestion des connaissances permet d'identifier et de capitaliser les savoirs faires de l'entreprise afin de les organiser et de les diffuser. Cette thèse propose un système de gestion des connaissances hétérogènes et distribuées, appelé OCEAN. Basé sur les ontologies et sur un système multi-agents, OCEAN a pour but de résoudre le problème de la capitalisation et de réutilisation des connaissances provenant de plusieurs sources différentes, afin d aider les acteurs métiers dans le processus de développement de produits mécaniques. Le système OCEAN repose sur un cycle de vie de quatre étapes Ce cycle de vie possède les phases : d identification, d extraction, de validation et se termine par la réutilisation des connaissances. Chaque phase constitue l objectif d une organisation d agents.L identification dans le système OCEAN consiste à définir les connaissances par un expert métier sous la forme d une ontologie. Les ontologies sont utilisées dans notre système pour représenter les connaissances définis d une façon structurée et formelle afin d être compréhensible par les machines. L extraction des connaissances dans OCEAN est réalisée par les agents de manière automatique à l aide des ontologies créées par les experts métiers. Les agents interagissent avec les différentes applications métiers via des services web. Le résultat de cette phase est stocké dans une mémoire organisationnelle. La validation des connaissances consiste à permettre aux acteurs métiers de valider les connaissances de la mémoire organisationnelle dans un wiki sémantique. Ce wiki permet de présenter les connaissances de la mémoire organisationnelle aux acteurs pour les réutiliser, les évaluer et les faire évoluer. La réutilisation des connaissances dans OCEAN est inspiré de travaux antérieurs intégrés au sein d OCEAN. Les quatre phases du cycle de vie des connaissances traitées dans cette thèse nous ont permis de réaliser un système apte à gérer les connaissances hétérogènes et distribuées dans une entreprise étendue.Among the goals of Knowledge Management we can cite the identification and capitalization of the know-how of companies in order to organize and disseminate them. This thesis proposes a heterogeneous and distributed knowledge management system, called OCEAN. Based on ontologies and multi-agents system, OCEAN aims to solve the problem of capitalization and reuse of multi-sources knowledge in order to assist business actors in the development process of mechanical products. The OCEAN system is based on a knowledge life cycle composed by four steps. This knowledge life cycle begins with the identification then extraction, validation and finishes with knowledge reuse. Each step is the goal of an organization of agents.The identification in OCEAN system consists in the definition of knowledge by a business expert with an ontology. Ontologies are used in our system to represent the knowledge, defined by the business expert, in a structured and formal way in order to be understandable by machines. Agents according to the ontology defined by business experts realize knowledge extraction in OCEAN automatically. Agents interact with professional softwares via web services. The result of this extraction is stored in an organizational memory (OM). Validation of knowledge in OCEAN relies on business actors that validate the knowledge of the OM in a semantic wiki. This wiki allows also the presentation of this knowledge to business actors in order to reuse, evaluate or evolve it. Previous works, integrated within OCEAN, inspires the knowledge reuse step. The four steps lifecycle discussed in this thesis has enabled us to achieve a system that can manage heterogeneous and distributed knowledge in an extended enterprise.BELFORT-UTBM-SEVENANS (900942101) / SudocSudocFranceF
Recommended from our members
Arabic Language Processing for Text Classification. Contributions to Arabic Root Extraction Techniques, Building An Arabic Corpus, and to Arabic Text Classification Techniques.
The impact and dynamics of Internet-based resources for Arabic-speaking users is increasing in significance, depth and breadth at highest pace than ever, and thus requires updated mechanisms for computational processing of Arabic texts. Arabic is a complex language and as such requires in depth investigation for analysis and improvement of available automatic processing techniques such as root extraction methods or text classification techniques, and for developing text collections that are already labeled, whether with single or multiple labels.
This thesis proposes new ideas and methods to improve available automatic processing techniques for Arabic texts. Any automatic processing technique would require data in order to be used and critically reviewed and assessed, and here an attempt to develop a labeled Arabic corpus is also proposed. This thesis is composed of three parts: 1- Arabic corpus development, 2- proposing, improving and implementing root extraction techniques, and 3- proposing and investigating the effect of different pre-processing methods on single-labeled text classification methods for Arabic.
This thesis first develops an Arabic corpus that is prepared to be used here for testing root extraction methods as well as single-label text classification techniques. It also enhances a rule-based root extraction method by handling irregular cases (that appear in about 34% of texts). It proposes and implements two expanded algorithms as well as an adjustment for a weight-based method. It also includes the algorithm that handles irregular cases to all and compares the performances of these proposed methods with original ones. This thesis thus develops a root extraction system that handles foreign Arabized words by constructing a list of about 7,000 foreign words. The outcome of the technique with best accuracy results in extracting the correct stem and root for respective words in texts, which is an enhanced rule-based method, is used in the third part of this thesis. This thesis finally proposes and implements a variant term frequency inverse document frequency weighting method, and investigates the effect of using different choices of features in document representation on single-label text classification performance (words, stems or roots as well as including to these choices their respective phrases). This thesis applies forty seven classifiers on all proposed representations and compares their performances. One challenge for researchers in Arabic text processing is that reported root extraction techniques in literature are either not accessible or require a long time to be reproduced while labeled benchmark Arabic text corpus is not fully available online. Also, by now few machine learning techniques were investigated on Arabic where usual preprocessing steps before classification were chosen. Such challenges are addressed in this thesis by developing a new labeled Arabic text corpus for extended applications of computational techniques.
Results of investigated issues here show that proposing and implementing an algorithm that handles irregular words in Arabic did improve the performance of all implemented root extraction techniques. The performance of the algorithm that handles such irregular cases is evaluated in terms of accuracy improvement and execution time. Its efficiency is investigated with different document lengths and empirically is found to be linear in time for document lengths less than about 8,000. The rule-based technique is improved the highest among implemented root extraction methods when including the irregular cases handling algorithm. This thesis validates that choosing roots or stems instead of words in documents representations indeed improves single-label classification performance significantly for most used classifiers. However, the effect of extending such representations with their respective phrases on single-label text classification performance shows that it has no significant improvement. Many classifiers were not yet tested for Arabic such as the ripple-down rule classifier. The outcome of comparing the classifiers' performances concludes that the Bayesian network classifier performance is significantly the best in terms of accuracy, training time, and root mean square error values for all proposed and implemented representations.Petra University, Amman (Jordan
International Academic Symposium of Social Science 2022
This conference proceedings gathers work and research presented at the International Academic Symposium of Social Science 2022 (IASSC2022) held on July 3, 2022, in Kota Bharu, Kelantan, Malaysia. The conference was jointly organized by the Faculty of Information Management of Universiti Teknologi MARA Kelantan Branch, Malaysia; University of Malaya, Malaysia; Universitas Pembangunan Nasional Veteran Jakarta, Indonesia; Universitas Ngudi Waluyo, Indonesia; Camarines Sur Polytechnic Colleges, Philippines; and UCSI University, Malaysia. Featuring experienced keynote speakers from Malaysia, Australia, and England, this proceeding provides an opportunity for researchers, postgraduate students, and industry practitioners to gain knowledge and understanding of advanced topics concerning digital transformations in the perspective of the social sciences and information systems, focusing on issues, challenges, impacts, and theoretical foundations. This conference proceedings will assist in shaping the future of the academy and industry by compiling state-of-the-art works and future trends in the digital transformation of the social sciences and the field of information systems. It is also considered an interactive platform that enables academicians, practitioners and students from various institutions and industries to collaborate