33 research outputs found

    Exploring Error-Sensitive Attributes and Branches of Decision Trees

    Full text link
    The 2013 9th International Conference on Natural Computation (ICNC'13) and the 2013 10th International Conference on Fuzzy Systems and Knowledge Discovery (FSKD'13) were be jointly held from 23-25 July 2013 in Shenyang, China. http://icnc-fskd.lntu.edu.cn

    Output Effect Evaluation Based on Input Features in Neural Incremental Attribute Learning for Better Classification Performance

    Get PDF
    [[abstract]]Machine learning is a very important approach to pattern classification. This paper provides a better insight into Incremental Attribute Learning (IAL) with further analysis as to why it can exhibit better performance than conventional batch training. IAL is a novel supervised machine learning strategy, which gradually trains features in one or more chunks. Previous research showed that IAL can obtain lower classification error rates than a conventional batch training approach. Yet the reason for that is still not very clear. In this study, the feasibility of IAL is verified by mathematical approaches. Moreover, experimental results derived by IAL neural networks on benchmarks also confirm the mathematical validation.[[notice]]補正完畢[[incitationindex]]SCI[[booktype]]電子

    Hybrid ant colony system algorithm for static and dynamic job scheduling in grid computing

    Get PDF
    Grid computing is a distributed system with heterogeneous infrastructures. Resource management system (RMS) is one of the most important components which has great influence on the grid computing performance. The main part of RMS is the scheduler algorithm which has the responsibility to map submitted tasks to available resources. The complexity of scheduling problem is considered as a nondeterministic polynomial complete (NP-complete) problem and therefore, an intelligent algorithm is required to achieve better scheduling solution. One of the prominent intelligent algorithms is ant colony system (ACS) which is implemented widely to solve various types of scheduling problems. However, ACS suffers from stagnation problem in medium and large size grid computing system. ACS is based on exploitation and exploration mechanisms where the exploitation is sufficient but the exploration has a deficiency. The exploration in ACS is based on a random approach without any strategy. This study proposed four hybrid algorithms between ACS, Genetic Algorithm (GA), and Tabu Search (TS) algorithms to enhance the ACS performance. The algorithms are ACS(GA), ACS+GA, ACS(TS), and ACS+TS. These proposed hybrid algorithms will enhance ACS in terms of exploration mechanism and solution refinement by implementing low and high levels hybridization of ACS, GA, and TS algorithms. The proposed algorithms were evaluated against twelve metaheuristic algorithms in static (expected time to compute model) and dynamic (distribution pattern) grid computing environments. A simulator called ExSim was developed to mimic the static and dynamic nature of the grid computing. Experimental results show that the proposed algorithms outperform ACS in terms of best makespan values. Performance of ACS(GA), ACS+GA, ACS(TS), and ACS+TS are better than ACS by 0.35%, 2.03%, 4.65% and 6.99% respectively for static environment. For dynamic environment, performance of ACS(GA), ACS+GA, ACS+TS, and ACS(TS) are better than ACS by 0.01%, 0.56%, 1.16%, and 1.26% respectively. The proposed algorithms can be used to schedule tasks in grid computing with better performance in terms of makespan

    Machine Learning in Chronic Pain Research: A Scoping Review

    Get PDF
    Given the high prevalence and associated cost of chronic pain, it has a significant impact on individuals and society. Improvements in the treatment and management of chronic pain may increase patients’ quality of life and reduce societal costs. In this paper, we evaluate state-of-the-art machine learning approaches in chronic pain research. A literature search was conducted using the PubMed, IEEE Xplore, and the Association of Computing Machinery (ACM) Digital Library databases. Relevant studies were identified by screening titles and abstracts for keywords related to chronic pain and machine learning, followed by analysing full texts. Two hundred and eighty-seven publications were identified in the literature search. In total, fifty-three papers on chronic pain research and machine learning were reviewed. The review showed that while many studies have emphasised machine learning-based classification for the diagnosis of chronic pain, far less attention has been paid to the treatment and management of chronic pain. More research is needed on machine learning approaches to the treatment, rehabilitation, and self-management of chronic pain. As with other chronic conditions, patient involvement and self-management are crucial. In order to achieve this, patients with chronic pain need digital tools that can help them make decisions about their own treatment and care

    EEG Eye State Identification Using Incremental Attribute Learning with Time-Series Classification

    Get PDF
    Eye state identification is a kind of common time-series classification problem which is also a hot spot in recent research. Electroencephalography (EEG) is widely used in eye state classification to detect human's cognition state. Previous research has validated the feasibility of machine learning and statistical approaches for EEG eye state classification. This paper aims to propose a novel approach for EEG eye state identification using incremental attribute learning (IAL) based on neural networks. IAL is a novel machine learning strategy which gradually imports and trains features one by one. Previous studies have verified that such an approach is applicable for solving a number of pattern recognition problems. However, in these previous works, little research on IAL focused on its application to time-series problems. Therefore, it is still unknown whether IAL can be employed to cope with time-series problems like EEG eye state classification. Experimental results in this study demonstrates that, with proper feature extraction and feature ordering, IAL can not only efficiently cope with time-series classification problems, but also exhibit better classification performance in terms of classification error rates in comparison with conventional and some other approaches

    A generalized laser simulator algorithm for optimal path planning in constraints environment

    Get PDF
    Path planning plays a vital role in autonomous mobile robot navigation, and it has thus become one of the most studied areas in robotics. Path planning refers to a robot's search for a collision-free and optimal path from a start point to a predefined goal position in a given environment. This research focuses on developing a novel path planning algorithm, called Generalized Laser Simulator (GLS), to solve the path planning problem of mobile robots in a constrained environment. This approach allows finding the path for a mobile robot while avoiding obstacles, searching for a goal, considering some constraints and finding an optimal path during the robot movement in both known and unknown environments. The feasible path is determined between the start and goal positions by generating a wave of points in all directions towards the goal point with adhering to constraints. A simulation study employing the proposed approach is applied to the grid map settings to determine a collision-free path from the start to goal positions. First, the grid mapping of the robot's workspace environment is constructed, and then the borders of the workspace environment are detected based on the new proposed function. This function guides the robot to move toward the desired goal. Two concepts have been implemented to find the best candidate point to move next: minimum distance to goal and maximum index distance to the boundary, integrated by negative probability to sort out the most preferred point for the robot trajectory determination. In order to construct an optimal collision-free path, an optimization step was included to find out the minimum distance within the candidate points that have been determined by GLS while adhering to particular constraint's rules and avoiding obstacles. The proposed algorithm will switch its working pattern based on the goal minimum and boundary maximum index distances. For static obstacle avoidance, the boundaries of the obstacle(s) are considered borders of the environment. However, the algorithm detects obstacles as a new border in dynamic obstacles once it occurs in front of the GLS waves. The proposed method has been tested in several test environments with different degrees of complexity. Twenty different arbitrary environments are categorized into four: Simple, complex, narrow, and maze, with five test environments in each. The results demonstrated that the proposed method could generate an optimal collision-free path. Moreover, the proposed algorithm result are compared to some common algorithms such as the A* algorithm, Probabilistic Road Map, RRT, Bi-directional RRT, and Laser Simulator algorithm to demonstrate its effectiveness. The suggested algorithm outperforms the competition in terms of improving path cost, smoothness, and search time. A statistical test was used to demonstrate the efficiency of the proposed algorithm over the compared methods. The GLS is 7.8 and 5.5 times faster than A* and LS, respectively, generating a path 1.2 and 1.5 times shorter than A* and LS. The mean value of the path cost achieved by the proposed approach is 4% and 15% lower than PRM and RRT, respectively. The mean path cost generated by the LS algorithm, on the other hand, is 14% higher than that generated by the PRM. Finally, to verify the performance of the developed method for generating a collision-free path, experimental studies were carried out using an existing WMR platform in labs and roads. The experimental work investigates complete autonomous WMR path planning in the lab and road environments using live video streaming. The local maps were built using data from live video streaming s by real-time image processing to detect the segments of the lab and road environments. The image processing includes several operations to apply GLS on the prepared local map. The proposed algorithm generates the path within the prepared local map to find the path between start-to-goal positions to avoid obstacles and adhere to constraints. The experimental test shows that the proposed method can generate the shortest path and best smooth trajectory from start to goal points in comparison with the laser simulator

    A local feature engineering strategy to improve network anomaly detection

    Get PDF
    The dramatic increase in devices and services that has characterized modern societies in recent decades, boosted by the exponential growth of ever faster network connections and the predominant use of wireless connection technologies, has materialized a very crucial challenge in terms of security. The anomaly-based intrusion detection systems, which for a long time have represented some of the most efficient solutions to detect intrusion attempts on a network, have to face this new and more complicated scenario. Well-known problems, such as the difficulty of distinguishing legitimate activities from illegitimate ones due to their similar characteristics and their high degree of heterogeneity, today have become even more complex, considering the increase in the network activity. After providing an extensive overview of the scenario under consideration, this work proposes a Local Feature Engineering (LFE) strategy aimed to face such problems through the adoption of a data preprocessing strategy that reduces the number of possible network event patterns, increasing at the same time their characterization. Unlike the canonical feature engineering approaches, which take into account the entire dataset, it operates locally in the feature space of each single event. The experiments conducted on real-world data showed that this strategy, which is based on the introduction of new features and the discretization of their values, improves the performance of the canonical state-of-the-art solutions

    Assembly sequence planning using hybrid binary particle swarm optimization

    Get PDF
    Assembly Sequence Planning (ASP) is known as a large-scale, timeconsuming combinatorial problem. Therefore time is the main factor in production planning. Recently, ASP in production planning had been studied widely especially to minimize the time and consequently reduce the cost. The first objective of this research is to formulate and analyse a mathematical model of the ASP problem. The second objective is to minimize the time of the ASP problem and hence reduce the product cost. A case study of a product consists of 19 components have been used in this research, and the fitness function of the problem had been calculated using Binary Particle Swarm Optimization (BPSO), and hybrid algorithm of BPSO and Differential Evolution (DE). The novel algorithm of BPSODE has been assessed with performance-evaluated criteria (performance measure). The algorithm has been validated using 8 comprehensive benchmark problems from the literature. The results show that the BPSO algorithm has an improved performance and can reduce further the time of assembly of the 19 parts of the ASP compared to the Simulated Annealing and Genetic Algorithm. The novel hybrid BPSODE algorithm shows a superior performance when assessed via performance-evaluated criteria compared to BPSO. The BPSODE algorithm also demonstrated a good generation of the recorded optimal value for the 8 standard benchmark problems

    A Review on the Application of Blockchain to the Next Generation of Cybersecure Industry 4.0 Smart Factories

    Get PDF
    [Absctract]: Industry 4.0 is a concept devised for improving the way modern factories operate through the use of some of the latest technologies, like the ones used for creating the Industrial Internet of Things (IIoT), robotics, or Big Data applications. One of such technologies is blockchain, which is able to add trust, security, and decentralization to different industrial fields. This article focuses on analyzing the benefits and challenges that arise when using blockchain and smart contracts to develop Industry 4.0 applications. In addition, this paper presents a thorough review of the most relevant blockchain-based applications for Industry 4.0 technologies. Thus, its aim is to provide a detailed guide for the future Industry 4.0 developers that allows for determining how the blockchain can enhance the next generation of cybersecure industrial applications
    corecore