2,384 research outputs found

    Profitable Task Allocation in Mobile Cloud Computing

    Full text link
    We propose a game theoretic framework for task allocation in mobile cloud computing that corresponds to offloading of compute tasks to a group of nearby mobile devices. Specifically, in our framework, a distributor node holds a multidimensional auction for allocating the tasks of a job among nearby mobile nodes based on their computational capabilities and also the cost of computation at these nodes, with the goal of reducing the overall job completion time. Our proposed auction also has the desired incentive compatibility property that ensures that mobile devices truthfully reveal their capabilities and costs and that those devices benefit from the task allocation. To deal with node mobility, we perform multiple auctions over adaptive time intervals. We develop a heuristic approach to dynamically find the best time intervals between auctions to minimize unnecessary auctions and the accompanying overheads. We evaluate our framework and methods using both real world and synthetic mobility traces. Our evaluation results show that our game theoretic framework improves the job completion time by a factor of 2-5 in comparison to the time taken for executing the job locally, while minimizing the number of auctions and the accompanying overheads. Our approach is also profitable for the nearby nodes that execute the distributor's tasks with these nodes receiving a compensation higher than their actual costs

    Optimal Order Assignment with Minimum Wage Consideration (OOAMWC)

    Get PDF
    While the application of crowdsourcing has increased over the years, the technology experiences various issues during implementation. Examples of some of the issues that affect crowdsourcing include task assignment, profit maximizations, as well as time window issues. In some instances addressing some of the issues results in the other issues being overlooked. An example is when assigning tasks to workers, the profits of the workers might not be considered and this ends up affecting the profit maximization aspect. Various algorithms have been proposed to address the task assignment, profit maximizations, and time window issues. However, these algorithms address the issues individually and this results in the occurrence of the other noted issues. Therefore, this calls for the definition of a solution to address the task assignment issue while taking into consideration the time window issue and the minimum wage constraint. Additionally, the solution should address the profit maximization of not only the workers but also the platform and the clients of the platform. To evaluate the efficiency of the proposed solution, a comparison with the different implemented solutions to address individual issues is recommended. Comparing such solutions can provide insight into the efficiency of the proposed approach when addressing multiple issues affecting crowdsourcing

    Certified Computation from Unreliable Datasets

    Full text link
    A wide range of learning tasks require human input in labeling massive data. The collected data though are usually low quality and contain inaccuracies and errors. As a result, modern science and business face the problem of learning from unreliable data sets. In this work, we provide a generic approach that is based on \textit{verification} of only few records of the data set to guarantee high quality learning outcomes for various optimization objectives. Our method, identifies small sets of critical records and verifies their validity. We show that many problems only need poly(1/ε)\text{poly}(1/\varepsilon) verifications, to ensure that the output of the computation is at most a factor of (1±ε)(1 \pm \varepsilon) away from the truth. For any given instance, we provide an \textit{instance optimal} solution that verifies the minimum possible number of records to approximately certify correctness. Then using this instance optimal formulation of the problem we prove our main result: "every function that satisfies some Lipschitz continuity condition can be certified with a small number of verifications". We show that the required Lipschitz continuity condition is satisfied even by some NP-complete problems, which illustrates the generality and importance of this theorem. In case this certification step fails, an invalid record will be identified. Removing these records and repeating until success, guarantees that the result will be accurate and will depend only on the verified records. Surprisingly, as we show, for several computation tasks more efficient methods are possible. These methods always guarantee that the produced result is not affected by the invalid records, since any invalid record that affects the output will be detected and verified
    • …
    corecore