31 research outputs found

    Efficient Approximation Algorithms for Multi-Antennae Largest Weight Data Retrieval

    Full text link
    In a mobile network, wireless data broadcast over mm channels (frequencies) is a powerful means for distributed dissemination of data to clients who access the channels through multi-antennae equipped on their mobile devices. The δ\delta-antennae largest weight data retrieval (δ\deltaALWDR) problem is to compute a schedule for downloading a subset of data items that has a maximum total weight using δ\delta antennae in a given time interval. In this paper, we propose a ratio 11eϵ1-\frac{1}{e}-\epsilon approximation algorithm for the δ\delta-antennae largest weight data retrieval (δ\deltaALWDR) problem that has the same ratio as the known result but a significantly improved time complexity of O(21ϵ1ϵm7T3.5L)O(2^{\frac{1}{\epsilon}}\frac{1}{\epsilon}m^{7}T^{3.5}L) from O(ϵ3.5m3.5ϵT3.5L)O(\epsilon^{3.5}m^{\frac{3.5}{\epsilon}}T^{3.5}L) when δ=1\delta=1 \cite{lu2014data}. To our knowledge, our algorithm represents the first ratio 11eϵ1-\frac{1}{e}-\epsilon approximation solution to δ\deltaALWDR for the general case of arbitrary δ\delta. To achieve this, we first give a ratio 11e1-\frac{1}{e} algorithm for the γ\gamma-separated δ\deltaALWDR (δ\deltaAγ\gammaLWDR) with runtime O(m7T3.5L)O(m^{7}T^{3.5}L), under the assumption that every data item appears at most once in each segment of δ\deltaAγ\gammaLWDR, for any input of maximum length LL on mm channels in TT time slots. Then, we show that we can retain the same ratio for δ\deltaAγ\gammaLWDR without this assumption at the cost of increased time complexity to O(2γm7T3.5L)O(2^{\gamma}m^{7}T^{3.5}L). This result immediately yields an approximation solution of same ratio and time complexity for δ\deltaALWDR, presenting a significant improvement of the known time complexity of ratio 11eϵ1-\frac{1}{e}-\epsilon approximation to the problem

    Efficient algorithm for the k-means problem with Must-Link and Cannot-Link constraints

    Get PDF
    Constrained clustering, such as k -means with instance-level Must-Link (ML) and Cannot-Link (CL) auxiliary information as the constraints, has been extensively studied recently, due to its broad applications in data science and AI. Despite some heuristic approaches, there has not been any algorithm providing a non-trivial approximation ratio to the constrained k -means problem. To address this issue, we propose an algorithm with a provable approximation ratio of O(logk) when only ML constraints are considered. We also empirically evaluate the performance of our algorithm on real-world datasets having artificial ML and disjoint CL constraints. The experimental results show that our algorithm outperforms the existing greedy-based heuristic methods in clustering accuracy

    Acceleration for Timing-Aware Gate-Level Logic Simulation with One-Pass GPU Parallelism

    Full text link
    Witnessing the advancing scale and complexity of chip design and benefiting from high-performance computation technologies, the simulation of Very Large Scale Integration (VLSI) Circuits imposes an increasing requirement for acceleration through parallel computing with GPU devices. However, the conventional parallel strategies do not fully align with modern GPU abilities, leading to new challenges in the parallelism of VLSI simulation when using GPU, despite some previous successful demonstrations of significant acceleration. In this paper, we propose a novel approach to accelerate 4-value logic timing-aware gate-level logic simulation using waveform-based GPU parallelism. Our approach utilizes a new strategy that can effectively handle the dependency between tasks during the parallelism, reducing the synchronization requirement between CPU and GPU when parallelizing the simulation on combinational circuits. This approach requires only one round of data transfer and hence achieves one-pass parallelism. Moreover, to overcome the difficulty within the adoption of our strategy in GPU devices, we design a series of data structures and tune them to dynamically allocate and store new-generated output with uncertain scale. Finally, experiments are carried out on industrial-scale open-source benchmarks to demonstrate the performance gain of our approach compared to several state-of-the-art baselines

    Improved approximation algorithms for computing kk k disjoint paths subject to two constraints

    Full text link

    On the shallow-light Steiner tree problem

    Full text link
    Let G = (V, E) be a given graph with nonnegative integral edge cost and delay, S ⊆ V be a terminal set and r ∈ S be the selected root. The shallow-light Steiner tree (SLST) problem is to compute a minimum cost tree spanning the terminals of S, such that the delay between r and every other terminal is bounded by a given delay constraint D ∈ ℤ 0 + . It is known that the SLST problem is NP-hard and unless NP ⊆ DTIME(n log log n ) there exists no approximation algorithm with ratio (1, γ log2 n) for some fixed γ > 0 [12]. Nevertheless, under the same assumption it admits no approximation ratio better than (1, γ log 2 n) for some fixed γ > 0 even when D = 2 [2]. This paper first gives an exact algorithm with time complexity O(3 t nD + 2 t n 2 D 2 + n 3 D 3 ), where n and t are the numbers of vertices and terminals of the given graph respectively. This is a pseudo polynomial time parameterized algorithm with respect to the parameterization “number of terminals”. Later, this algorithm is improved to a parameterized approximation algorithm with a time complexity O(3 t n 2 /∈ + 2 t n 4 /∈ 2 + n 6 /∈ 3 ) and a bifactor approximation ratio (1 + ∈, 1). That is, for any small real number ∈ > 0, the algorithm computes a Steiner tree with delay and cost bounded by (1 + ∈)D and the optimum cost respectively
    corecore