30,494 research outputs found
Scheduling with Outliers
In classical scheduling problems, we are given jobs and machines, and have to
schedule all the jobs to minimize some objective function. What if each job has
a specified profit, and we are no longer required to process all jobs -- we can
schedule any subset of jobs whose total profit is at least a (hard) target
profit requirement, while still approximately minimizing the objective
function?
We refer to this class of problems as scheduling with outliers. This model
was initiated by Charikar and Khuller (SODA'06) on the minimum max-response
time in broadcast scheduling. We consider three other well-studied scheduling
objectives: the generalized assignment problem, average weighted completion
time, and average flow time, and provide LP-based approximation algorithms for
them. For the minimum average flow time problem on identical machines, we give
a logarithmic approximation algorithm for the case of unit profits based on
rounding an LP relaxation; we also show a matching integrality gap. For the
average weighted completion time problem on unrelated machines, we give a
constant factor approximation. The algorithm is based on randomized rounding of
the time-indexed LP relaxation strengthened by the knapsack-cover inequalities.
For the generalized assignment problem with outliers, we give a simple
reduction to GAP without outliers to obtain an algorithm whose makespan is
within 3 times the optimum makespan, and whose cost is at most (1 + \epsilon)
times the optimal cost.Comment: 23 pages, 3 figure
An Exploratory Study of Patient Falls
Debate continues between the contribution of education level and clinical expertise in the nursing practice environment. Research suggests a link between Baccalaureate of Science in Nursing (BSN) nurses and positive patient outcomes such as lower mortality, decreased falls, and fewer medication errors. Purpose: To examine if there a negative correlation between patient falls and the level of nurse education at an urban hospital located in Midwest Illinois during the years 2010-2014? Methods: A retrospective crosssectional cohort analysis was conducted using data from the National Database of Nursing Quality Indicators (NDNQI) from the years 2010-2014. Sample: Inpatients aged ≥ 18 years who experienced a unintentional sudden descent, with or without injury that resulted in the patient striking the floor or object and occurred on inpatient nursing units. Results: The regression model was constructed with annual patient falls as the dependent variable and formal education and a log transformed variable for percentage of certified nurses as the independent variables. The model overall is a good fit, F (2,22) = 9.014, p = .001, adj. R2 = .40. Conclusion: Annual patient falls will decrease by increasing the number of nurses with baccalaureate degrees and/or certifications from a professional nursing board-governing body
Splitting hybrid Make-To-Order and Make-To-Stock demand profiles
In this paper a demand time series is analysed to support Make-To-Stock (MTS)
and Make-To-Order (MTO) production decisions. Using a purely MTS production
strategy based on the given demand can lead to unnecessarily high inventory
levels thus it is necessary to identify likely MTO episodes.
This research proposes a novel outlier detection algorithm based on special
density measures. We divide the time series' histogram into three clusters. One
with frequent-low volume covers MTS items whilst a second accounts for high
volumes which is dedicated to MTO items. The third cluster resides between the
previous two with its elements being assigned to either the MTO or MTS class.
The algorithm can be applied to a variety of time series such as stationary and
non-stationary ones.
We use empirical data from manufacturing to study the extent of inventory
savings. The percentage of MTO items is reflected in the inventory savings
which were shown to be an average of 18.1%.Comment: demand analysis; time series; outlier detection; production strategy;
Make-To-Order(MTO); Make-To-Stock(MTS); 15 pages, 9 figure
Energy Efficient Scheduling via Partial Shutdown
Motivated by issues of saving energy in data centers we define a collection
of new problems referred to as "machine activation" problems. The central
framework we introduce considers a collection of machines (unrelated or
related) with each machine having an {\em activation cost} of . There
is also a collection of jobs that need to be performed, and is
the processing time of job on machine . We assume that there is an
activation cost budget of -- we would like to {\em select} a subset of
the machines to activate with total cost and {\em find} a schedule
for the jobs on the machines in minimizing the makespan (or any other
metric).
For the general unrelated machine activation problem, our main results are
that if there is a schedule with makespan and activation cost then we
can obtain a schedule with makespan \makespanconstant T and activation cost
\costconstant A, for any . We also consider assignment costs for
jobs as in the generalized assignment problem, and using our framework, provide
algorithms that minimize the machine activation and the assignment cost
simultaneously. In addition, we present a greedy algorithm which only works for
the basic version and yields a makespan of and an activation cost .
For the uniformly related parallel machine scheduling problem, we develop a
polynomial time approximation scheme that outputs a schedule with the property
that the activation cost of the subset of machines is at most and the
makespan is at most for any
Autonomic care platform for optimizing query performance
Background: As the amount of information in electronic health care systems increases, data operations get more complicated and time-consuming. Intensive Care platforms require a timely processing of data retrievals to guarantee the continuous display of recent data of patients. Physicians and nurses rely on this data for their decision making. Manual optimization of query executions has become difficult to handle due to the increased amount of queries across multiple sources. Hence, a more automated management is necessary to increase the performance of database queries. The autonomic computing paradigm promises an approach in which the system adapts itself and acts as self-managing entity, thereby limiting human interventions and taking actions. Despite the usage of autonomic control loops in network and software systems, this approach has not been applied so far for health information systems.
Methods: We extend the COSARA architecture, an infection surveillance and antibiotic management service platform for the Intensive Care Unit (ICU), with self-managed components to increase the performance of data retrievals. We used real-life ICU COSARA queries to analyse slow performance and measure the impact of optimizations. Each day more than 2 million COSARA queries are executed. Three control loops, which monitor the executions and take action, have been proposed: reactive, deliberative and reflective control loops. We focus on improvements of the execution time of microbiology queries directly related to the visual displays of patients' data on the bedside screens.
Results: The results show that autonomic control loops are beneficial for the optimizations in the data executions in the ICU. The application of reactive control loop results in a reduction of 8.61% of the average execution time of microbiology results. The combined application of the reactive and deliberative control loop results in an average query time reduction of 10.92% and the combination of reactive, deliberative and reflective control loops provides a reduction of 13.04%.
Conclusions: We found that by controlled reduction of queries' executions the performance for the end-user can be improved. The implementation of autonomic control loops in an existing health platform, COSARA, has a positive effect on the timely data visualization for the physician and nurse
- …