1,125 research outputs found

    Bi-velocity discrete particle swarm optimization and its application to multicast routing problem in communication networks

    Get PDF
    This paper proposes a novel bi-velocity discrete particle swarm optimization (BVDPSO) approach and extends its application to the NP-complete multicast routing problem (MRP). The main contribution is the extension of PSO from continuous domain to the binary or discrete domain. Firstly, a novel bi-velocity strategy is developed to represent possibilities of each dimension being 1 and 0. This strategy is suitable to describe the binary characteristic of the MRP where 1 stands for a node being selected to construct the multicast tree while 0 stands for being otherwise. Secondly, BVDPSO updates the velocity and position according to the learning mechanism of the original PSO in continuous domain. This maintains the fast convergence speed and global search ability of the original PSO. Experiments are comprehensively conducted on all of the 58 instances with small, medium, and large scales in the OR-library (Operation Research Library). The results confirm that BVDPSO can obtain optimal or near-optimal solutions rapidly as it only needs to generate a few multicast trees. BVDPSO outperforms not only several state-of-the-art and recent heuristic algorithms for the MRP problems, but also algorithms based on GA, ACO, and PSO

    The Application of Ant Colony Optimization

    Get PDF
    The application of advanced analytics in science and technology is rapidly expanding, and developing optimization technics is critical to this expansion. Instead of relying on dated procedures, researchers can reap greater rewards by utilizing cutting-edge optimization techniques like population-based metaheuristic models, which can quickly generate a solution with acceptable quality. Ant Colony Optimization (ACO) is one the most critical and widely used models among heuristics and meta-heuristics. This book discusses ACO applications in Hybrid Electric Vehicles (HEVs), multi-robot systems, wireless multi-hop networks, and preventive, predictive maintenance

    A Study Resource Optimization Techniques Based Job Scheduling in Cloud Computing

    Get PDF
    Cloud computing has revolutionized the way businesses and individuals utilize computing resources. It offers on-demand access to a vast pool of virtualized resources, such as processing power, storage, and networking, through the Internet. One of the key challenges in cloud computing is efficiently scheduling jobs to maximize resource utilization and minimize costs. Job scheduling in cloud computing involves allocating tasks or jobs to available resources in an optimal manner. The objective is to minimize job completion time, maximize resource utilization, and meet various performance metrics such as response time, throughput, and energy consumption. Resource optimization techniques play a crucial role in achieving these objectives. Resource optimization techniques aim to efficiently allocate resources to jobs, taking into account factors like resource availability, job priorities, and constraints. These techniques utilize various algorithms and optimization approaches to make intelligent decisions about resource allocation. Research on resource optimization techniques for job scheduling in cloud computing is of significant importance due to the following reasons: Efficient Resource Utilization: Cloud computing environments consist of a large number of resources that need to be utilized effectively to maximize cost savings and overall system performance. By optimizing job scheduling, researchers can develop algorithms and techniques that ensure efficient utilization of resources, leading to improved productivity and reduced costs. Performance Improvement: Job scheduling plays a crucial role in meeting performance metrics such as response time, throughput, and reliability. By designing intelligent scheduling algorithms, researchers can improve the overall system performance, leading to better user experience and customer satisfaction. Scalability: Cloud computing environments are highly scalable, allowing users to dynamically scale resources based on their needs. Effective job scheduling techniques enable efficient resource allocation and scaling, ensuring that the system can handle varying workloads without compromising performance. Energy Efficiency: Cloud data centres consume significant amounts of energy, and optimizing resource allocation can contribute to energy conservation. By scheduling jobs intelligently, researchers can reduce energy consumption, leading to environmental benefits and cost savings for cloud service providers. Quality of Service (QoS): Cloud computing service providers often have service-level agreements (SLAs) that define the QoS requirements expected by users. Resource optimization techniques for job scheduling can help meet these SLAs by ensuring that jobs are allocated resources in a timely manner, meeting performance guarantees, and maintaining high service availability. Here in this research, we have used the method of the weighted product model (WPM). For the topic of Resource Optimization Techniques Based Job Scheduling in Cloud Computing For calculating the values of alternative and evaluation parameters. A variation of the WSM called the weighted product method (WPM) has been proposed to address some of the weaknesses of The WSM that came before it. The main distinction is that the multiplication is being used in place of additional. The terms "scoring methods" are frequently used to describe WSM and WPM Execution time on Virtual machine, Transmission time (delay)on Virtual machine, Processing cost of a task on virtual machine resource optimization techniques based on job scheduling play a crucial role in maximizing the efficiency and performance of cloud computing systems. By effectively managing and allocating resources, these techniques help minimize costs, reduce energy consumption, and improve overall system throughput. One of the key findings is that intelligent job scheduling algorithms, such as genetic algorithms, ant colony optimization

    Performance analysis for network coding using ant colony routing

    Get PDF
    This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The aim of this thesis is to conduct performance investigation of a combined system of Network Coding (NC) technique with Ant-Colony (ACO) routing protocol. This research analyses the impact of several workload characteristics, on system performance. Network coding is a significant key development of information transmission and processing. Network coding enhances the performance of multicast by employing encoding operations at intermediate nodes. Two steps should realize while using network coding in multicast communication: determining appropriate transmission paths from source to multi-receivers and using the suitable coding scheme. Intermediate nodes would combine several packets and relay them as a single packet. Although network coding can make a network achieve the maximum multicast rate, it always brings additional overheads. It is necessary to minimize unneeded overhead by using an optimization technique. On other hand, Ant Colony Optimization can be transformed into useful technique that seeks imitate the ant’s behaviour in finding the shortest path to its destination using quantities of pheromone that is left by former ants as guidance, so by using the same concept of the communication network environment, shorter paths can be formulated. The simulation results show that the resultant system considerably improves the performance of the network, by combining Ant Colony Optimization with network coding. 25% improvement in the bandwidth consumption can be achieved in comparison with conventional routing protocols. Additionally simulation results indicate that the proposed algorithm can decrease the computation time of system by a factor of 20%

    A green intelligent routing algorithm supporting flexible QoS for many-to-many multicast

    Get PDF
    The tremendous energy consumption attributed to the Information and Communication Technology (ICT) field has become a persistent concern during the last few years, attracting significant academic and industrial efforts. Networks have begun to be improved towards being “green”. Considering Quality of Service (QoS) and power consumption for green Internet, a Green Intelligent flexible QoS many-to-many Multicast routing algorithm (GIQM) is presented in this paper. In the proposed algorithm, a Rendezvous Point Confirming Stage (RPCS) is first carried out to obtain a rendezvous point and the candidate Many-to-many Multicast Sharing Tree (M2ST); then an Optimal Solution Identifying Stage (OSIS) is performed to generate a modified M2ST rooted at the rendezvous point, and an optimal M2ST is obtained by comparing the original M2ST and the modified M2ST. The network topology of Cernet2, GéANT and Internet2 were considered for the simulation of GIQM. The results from a series of experiments demonstrate the good performance and outstanding power-saving potential of the proposed GIQM with QoS satisfied

    Achieving Energy Efficiency on Networking Systems with Optimization Algorithms and Compressed Data Structures

    Get PDF
    To cope with the increasing quantity, capacity and energy consumption of transmission and routing equipment in the Internet, energy efficiency of communication networks has attracted more and more attention from researchers around the world. In this dissertation, we proposed three methodologies to achieve energy efficiency on networking devices: the NP-complete problems and heuristics, the compressed data structures, and the combination of the first two methods. We first consider the problem of achieving energy efficiency in Data Center Networks (DCN). We generalize the energy efficiency networking problem in data centers as optimal flow assignment problems, which is NP-complete, and then propose a heuristic called CARPO, a correlation-aware power optimization algorithm, that dynamically consolidate traffic flows onto a small set of links and switches in a DCN and then shut down unused network devices for power savings. We then achieve energy efficiency on Internet routers by using the compressive data structure. A novel data structure called the Probabilistic Bloom Filter (PBF), which extends the classical bloom filter into the probabilistic direction, so that it can effectively identify heavy hitters with a small memory foot print to reduce energy consumption of network measurement. To achieve energy efficiency on Wireless Sensor Networks (WSN), we developed one data collection protocol called EDAL, which stands for Energy-efficient Delay-aware Lifetime-balancing data collection. Based on the Open Vehicle Routing problem, EDAL exploits the topology requirements of Compressive Sensing (CS), then implement CS to save more energy on sensor nodes

    Synergy between biology and systems resilience

    Get PDF
    Resilient systems have the ability to endure and successfully recover from disturbances by identifying problems and mobilizing the available resources to cope with the disturbance. Resiliency lets a system recover from disruptions, variations, and a degradation of expected working conditions. Biological systems are resilient. Immune systems are highly adaptive and scalable, with the ability to cope with multiple data sources, fuse information together, makes decisions, have multiple interacting agents, operate in a distributed manner over a multiple scales, and have a memory structure to facilitate learning. Ecosystems are resilient since they have the capacity to absorb disturbance and are able to tolerate the disturbances. Ants build colonies that are dispersed, modular, fine grained, and standardized in design, yet they manage to forage intelligently for food and also organize collective defenses by the property of resilience. Are there any rules that we can identify to explain the resilience in these systems? The answer is yes. In insect colonies, rules determine the division of labor and how individual insects act towards each other and respond to different environmental possibilities. It is possible to group these rules based on attributes. These attributes are distributability, redundancy, adaptability, flexibility, interoperability, and diversity. It is also possible to incorporate these rules into engineering systems in their design to make them resilient. It is also possible to develop a qualitative model to generate resilience heuristics for engineering system based on a given attribute. The rules seen in nature and those of an engineering system are integrated to incorporate the desired characteristics for system resilience. The qualitative model for systems resilience will be able to generate system resilience heuristics. This model is simple and it can be applied to any system by using attribute based heuristics that are domain dependent. It also provides basic foundation for building computational models for designing resilient system architectures. This model was tested on recent catastrophes like the Mumbai terror attack and hurricane Katrina. With the disturbances surrounding the current world this resilience model based on heuristics will help a system to deal with crisis and still function in the best way possible by depending mainly on internal variables within the system --Abstract, page iii
    • …
    corecore