12,513 research outputs found

    Effects of partitioning and scheduling sparse matrix factorization on communication and load balance

    Get PDF
    A block based, automatic partitioning and scheduling methodology is presented for sparse matrix factorization on distributed memory systems. Using experimental results, this technique is analyzed for communication and load imbalance overhead. To study the performance effects, these overheads were compared with those obtained from a straightforward 'wrap mapped' column assignment scheme. All experimental results were obtained using test sparse matrices from the Harwell-Boeing data set. The results show that there is a communication and load balance tradeoff. The block based method results in lower communication cost whereas the wrap mapped scheme gives better load balance

    Load balance algorithms for anycast

    Full text link
    Increasingly, replicated anycast servers are being used to deliver network applications and service ever increasing user requests. Therefore, the strategies used to guarantee network bandwidth prerequisites and perform load balancing across the nodes of an anycast group are critical to the performance of online applications. In this paper, we model user requests, network congestion and latency, and server load using a combination of hydro-dynamics and queuing theory to develop an efficient job distribution strategy. Current, anycast research does not explicitly consider the system load of nodes within an anycast groups when distributing requests. Therefore, the performance of a heavily loaded anycast system can quickly become congested and uneven as jobs are routed to closely linked nodes which are already saturated with requests. In comparison, the nodes of further away systems remain relatively unused because of other issues such as network bandwidth and latency during these times. Our system redirects requests from busy systems to the idle, remotely linked nodes, to process requests faster in spite of slower network access. Using an empirical study, we show this technique can improve request performance, and throughput with minimal network probing overhead.<br /
    corecore