12
* Correspondence to: Gerhard J. Woeginger, Institut fu¨r Mathematik, Tu Graz, Steyrergasse 30, A-8010 Graz, Austria. E-mail: gwoegi@igi.tu-graz.ac.at Contract/grant sponsor: USAIsraeli BSF Contract/grant sponsor: Austrian Ministry of Science CCC 1094 6136/98/010055 12$17.50 ( 1998 John Wiley & Sons, Ltd. JOURNAL OF SCHEDULING J. Sched. 1, 55 66 (1998) APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES NOGA ALON1, YOSSI AZAR2, GERHARD J. WOEGINGER3,* AND TAL YADID4 1 School of Mathematical Sciences, ¹ el-Aviv ºniversity, ¹ el-Aviv, Israel 2 Department of Computer Science, ¹ el-Aviv ºniversity, ¹ el-Aviv, Israel 3 Institut fu ¨ r Mathematik, ¹º Graz, Steyrergasse 30, A-8010 Graz, Austria 4 Department of Computer Science, ¹ el-Aviv ºniversity, ¹ el-Aviv, Israel ABSTRACT We discuss scheduling problems with m identical machines and n jobs where each job has to be assigned to some machine. The goal is to optimize objective functions that solely depend on the machine completion times. As a main result, we identify some conditions on the objective function, under which the resulting scheduling problems possess a polynomial-time approximation scheme. Our result contains, generalizes, improves, simplifies, and unifies many other results in this area in a natural way. ( 1998 John Wiley & Sons, Ltd. KEY WORDS: scheduling theory; approximation algorithm; approximation scheme; worst-case ratio; combinatorial optimization 1. INTRODUCTION In this paper we consider scheduling problems with m*2 identical machines M i ,1)i)m, and n independent jobs J j ,1)j)n, where job J j has a positive processing time (or length) p j . A schedule is an assignment of the n jobs to the m machines. For 1)i)m, the completion time C i of machine M i in some fixed schedule equals the total processing time of the jobs that are assigned to M i . The goal is to find a schedule that optimizes an objective function that solely depends on the machine completion times. More precisely, for a fixed function f : R0 PR0 , we consider the problems of (I) minimizing the objective function + m i/1 f (C i ), (II) minimizing the objective function maxm i/1 f (C i ). In the standard scheduling notation of Lawler et al. [1], these problems are denoted by P D ) D + f (C i ) and by P D ) D max f (C i ), respectively. Moreover, we will deal with maximization problems that are dual to the problems (I) and (II), i.e. with the problem (I@) of maximizing the objective function + m i/1 f (C i ) and with the problem (II@) of maximizing minm i/1 f (C i ).

Approximation schemes for scheduling on parallel machines

Embed Size (px)

Citation preview

Page 1: Approximation schemes for scheduling on parallel machines

*Correspondence to: Gerhard J. Woeginger, Institut fur Mathematik, Tu Graz, Steyrergasse 30, A-8010 Graz, Austria.E-mail: [email protected]

Contract/grant sponsor: USA—Israeli BSFContract/grant sponsor: Austrian Ministry of Science

CCC 1094—6136/98/010055—12$17.50( 1998 John Wiley & Sons, Ltd.

JOURNAL OF SCHEDULING

J. Sched. 1, 55—66 (1998)

APPROXIMATION SCHEMES FOR SCHEDULINGON PARALLEL MACHINES

NOGA ALON1, YOSSI AZAR2, GERHARD J. WOEGINGER3,* AND TAL YADID4

1 School of Mathematical Sciences, ¹el-Aviv ºniversity, ¹el-Aviv, Israel2Department of Computer Science, ¹el-Aviv ºniversity, ¹el-Aviv, Israel

3 Institut f ur Mathematik, ¹º Graz, Steyrergasse 30, A-8010 Graz, Austria4Department of Computer Science, ¹el-Aviv ºniversity, ¹el-Aviv, Israel

ABSTRACT

We discuss scheduling problems with m identical machines and n jobs where each job has to be assigned tosome machine. The goal is to optimize objective functions that solely depend on the machine completion times.

As a main result, we identify some conditions on the objective function, under which the resultingscheduling problems possess a polynomial-time approximation scheme. Our result contains, generalizes,improves, simplifies, and unifies many other results in this area in a natural way. ( 1998 John Wiley& Sons, Ltd.

KEY WORDS: scheduling theory; approximation algorithm; approximation scheme; worst-case ratio; combinatorialoptimization

1. INTRODUCTION

In this paper we consider scheduling problems with m*2 identical machines Mi, 1)i)m, and

n independent jobs Jj, 1)j)n, where job J

jhas a positive processing time (or length) p

j.

A schedule is an assignment of the n jobs to the m machines. For 1)i)m, the completion timeC

iof machine M

iin some fixed schedule equals the total processing time of the jobs that are

assigned to Mi. The goal is to find a schedule that optimizes an objective function that solely

depends on the machine completion times. More precisely, for a fixed function f : R`0PR`

0, we

consider the problems of

(I) minimizing the objective function +mi/1

f (Ci),

(II) minimizing the objective function maxmi/1

f (Ci).

In the standard scheduling notation of Lawler et al. [1], these problems are denoted byP D ) D + f (C

i) and by P D ) Dmax f (C

i), respectively. Moreover, we will deal with maximization

problems that are dual to the problems (I) and (II), i.e. with the problem (I@) of maximizing theobjective function + m

i/1f (C

i) and with the problem (II@) of maximizing minm

i/1f (C

i).

Page 2: Approximation schemes for scheduling on parallel machines

All non-trivial scheduling problems of the forms (I), (II), (I@), and (II@) that have been discussed inthe literature are NP-complete in the strong sense. Hence, research focused on obtainingpolynomial-time approximation algorithms for this problem, i.e. fast algorithms which constructschedules whose objective function value is not too far away from the optimum value. We say thatan approximation algorithm for a minimization problem has performance ratio or performanceguarantee o for some real o'1, if it always delivers a solution with objective function value at mosto times the optimum value. Such an approximation algorithm is then called a o-approximationalgorithm. A family of polynomial-time (1#e)-approximation algorithms over all e'0 is calleda polynomial-time approximation scheme or PTAS, for short. For maximization problems, o-approxi-mation algorithms with o(1 are defined in a symmetric way. Moreover, a PTAS for a maximizationproblem is a family of polynomial-time (1!e)-approximation algorithms over all e'0.

1.1. Four examples

Let us give four illustrating examples for scheduling problems with objective functions of thetype (I), (II), and (II@):

Example 1. Minimizing the maximum machine completion time, i.e. problem (II) withf (x)"x. This definitely is the most classical scheduling problem on parallel machines. Already inthe 1960s, Graham [2, 3] analysed the worst case behaviour of variants of the greedy algorithmfor this problem: ¸ist Scheduling (LS) starts with an arbitrarily ordered list of jobs and repeatedlyassigns the next job in the list to a machine with currently smallest load. The ¸ongest Processing¹ime rule (LPT) first sorts the jobs by non-increasing processing time and then behaves like LS.Graham proved that LS has performance guarantee 2 and that LPT has performance guarantee4/3. Sahni [4] constructed a PTAS for the case where the number of machines is not part of theinput. Finally, Hochbaum and Shmoys [5] completely settled the approximability status of thisproblem by constructing a PTAS for the most general case with an arbitrary number of machines.

Example 2. Minimizing the sum of squares of the machine completion times, i.e. problem (I)with f (x)"x2. This problem arises e.g. when placing a set of records on a sectored drum so as tominimize the average latency-time [6] and in other storage allocation problems [7]. Chandra andWong [7] analysed the LPT rule for this problem and proved a performance guarantee of 25/24;this result was slightly improved by Leung and Wei [8]. Avidor, et al. [9] showed that LS hasperformance guarantee 4/3.

A more general problem is to minimize the sum of the p-th powers of the machine completiontimes, i.e. problem (I) with f (x)"xp . Note that this problem is equivalent to minimizing the¸p-norm of the vector (C

1, . . . , C

m), i.e. to minimize (+m

i/1Cp

i)1@p . Chandra and Wong [7]

showed that for every fixed p*1, the LPT rule achieves a constant performance ratio whosevalue depends on p and may be as large as 3/2. Alon et al. [10] finally exhibited a PTAS for everyfixed p*1. Their approach exploits the equivalence with the minimization of the ¸

p-norm and

relies heavily on the triangle inequality.

Example 3. Scheduling machines with extensible working times, i.e. problem (I) withf (x)"maxM1, xN. This problem arises e.g. in systems with m workers who have regular workingtimes of one time unit and who may make (if needed) extra-paid overtime work. The goal is toassign duties to the workers so as to minimize the total cost, i.e. the fixed cost for regular workingtimes plus the extra cost for overtime work. Dell’Olmo et al. [11] showed that the LPT rule has

56 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 3: Approximation schemes for scheduling on parallel machines

a performance guarantee of 13/11. A straightforward modification of the Hochbaum and Shmoys[5] approach yields a PTAS for this problem.

Example 4. Maximizing the minimum machine completion time, i.e. problem (II@) withf (x)"x. This problem has applications in the sequencing of maintenance actions for modular gasturbine aircraft engines, see Friesen and Deuermeyer [12]. Deuermeyer et al. [13] showed thatthe LPT rule has a performance guarantee of 3/4; Csirik et al. [14] tightened this analysis. Finally,Woeginger [15] gave a PTAS for this problem.

1.2. Results of this paper

We identify some conditions on functions f for which the corresponding scheduling problems(I), (II), (I@), and (II@) possess a polynomial time approximation scheme. Function f : R`

0PR`

0is

convex if and only if f (x#*)#f (y!*))f (x)#f (y) holds for all 0)x)y with 0)*)

y!x. Function f is concave if and only if f (x#*)#f (y!*)*f (x)#f (y) holds for all0)x)y with 0)*)y!x. The following condition (F*) will be essential throughout thepaper.

(F*): For all e'0 there exists a d'0 whose value depends only on e, such that:

∀x, y*0, (1!d) x)y)(1#d) x N (1!e) f (x))f (y))(1#e) f (x)

Condition (F*) essentially requires that the function f is continuous in a very strong way. Thecondition (F*) is discussed in detail in Section 4. Our main result on problems (I) and (II) is statedas follows.

¹heorem 1.1. For every non-negative and convex function f that fulfills condition (F*), thescheduling problem of minimizing + f (C

i) and the scheduling problem of minimizing max f (C

i) both

possess polynomial time approximation schemes. ¹he running time of these P¹AS is O(n), where thehidden constant depends exponentially on the reciprocal value of the desired precision.

Note that the time complexity of this PTAS is best possible with respect to the number of jobs.Moreover, the existence of a PTAS whose running time is also polynomial in the reciprocal valueof the precision for a strongly NP-complete problem would imply P"NP; cf. p. 141 of Garey andJohnson [16].

An argument that is completely symmetric to the proof of Theorem 1.1 will yield a correspond-ing theorem for the maximization problems (I@) and (II@).

¹heorem 1.2. For every non-negative concave function f that fulfills condition (F*), the schedulingproblem of maximizing + f (C

i) and the scheduling problem of maximizing min f (C

i) both possess

a polynomial-time approximation scheme with linear running time.

The reader may want to verify that the functions f (x)"x, f (x)"xp , and f (x)"maxM1, xN asintroduced in Examples 1—3 above are convex and satisfy condition (F*), and that the functionf (x)"x in Example 4 is concave and fulfills condition (F*). Hence, our Theorems 1.1 and 1.2immediately imply the existence of a PTAS for the scheduling problems in Examples 1—4.

Our approximation scheme generalizes the schemes of Sahni [4] and of Hochbaum andShmoys [5], and in fact it strongly builds on their ideas. However, there are several differenceswhere a major one is dealing with the so-called small jobs, i.e. jobs whose processing times aresmaller than a certain data-dependent threshold. The algorithms in [4, 5] first remove these small

APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES 57

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 4: Approximation schemes for scheduling on parallel machines

jobs, then solve the remaining problem to get an intermediate auxiliary schedule, and at lastgreedily add the small jobs to the intermediate schedule. A similar approach provably does nothelp in proving Theorems 1.1 and 1.2 (see Example 3.1 in Section 3). Instead, we will find a way ofmerging the small jobs into packages of reasonable size and then work on with these packages.

Organization of the paperIn Section 2 we construct a polynomial-time approximation scheme for scheduling problems (I)

where the non-negative, convex function f fulfills property (F*). Section 3 shows how to extendthese results to scheduling problems (II), (I@), and (II@). Section 4 discusses property (F*), and itinvestigates the relationship between the existence of a PTAS and the worst-case behaviour of theLPT rule. Finally, Section 5 contains the conclusion.

2. A GENERIC POLYNOMIAL-TIME APPROXIMATION SCHEME

Throughout this section, we will deal with an instance I with m machines, n jobs with jobprocessing times p

1, . . . , p

n, and the objective of minimizing +m

i/1f (C

i) where the non-negative

convex function f satisfies (F*). Our first goal will be to prove some simple combinatorialobservations. Let

¸"

1

m

n+j/1

pj

(1)

denote the average load of the m machines.

Observation 2.1. If a job Jjhas length p

j*¸, then there exists an optimum schedule in which job

Jjis the only job assigned to its machine.

Proof. Suppose that in some optimum schedule, job Jjis assigned together with some other

jobs to machine M1. Then the completion time C

1of machine M

1fulfills C

1"p

j#* with *'0.

Moreover, there exists a machine, say M2, whose completion time C

2is strictly less than ¸. But

now since *)C1!C

2then

f (C1)#f (C

2)*f (C

1!*)#f (C

2#*) (2)

holds by the convexity of the function f. Hence, moving all jobs except Jjfrom M

1to M

2cannot

increase the objective function. j

With the help of Observation 2.1 we may iteratively assign the very large jobs whose lengthis *¸ : While there is a job whose length is larger than the current average load, we assign thisjob to a machine and remove both the job and the machine from the setting. Note that in doingthis, both the optimal algorithm and the iterative process assign the large jobs to separatemachines. Thus, we may assume without loss of generality that in the end no job has length largerthan the average load. More specifically, from a given PTAS for the case where none of the jobshas length larger than the average, we can generate a PTAS for the general case, by assigning thelarge jobs to separate machines and applying the PTAS on the remaining jobs and machines.

Hence, from now on we may (and will) assume that no job has processing time greater than orequal to ¸.

Observation 2.2. If no job has processing time greater than or equal to ¸, then there exists anoptimum schedule in which all machine completion times C

i, 1)i)m, fulfill 1

2¸(C

i(2¸.

58 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 5: Approximation schemes for scheduling on parallel machines

Proof. Among all optimum schedules, consider a schedule that minimizes the auxiliary func-tion +m

i/1C2

i. We claim that in this optimum schedule 1

2¸(C

i(2¸ holds for all machine

completion times.First suppose that C

i*2¸ holds for some machine M

i. Similarly as in the proof of Observa-

tion 2.1, moving an arbitrary job from Mito the machine with smallest completion time cannot

increase the convex objective function, but does decrease the auxiliary function; a contradiction.Next, suppose that C

i)1

2¸ holds for some machine M

i. Consider another machine M

jwith

Cj'¸. In case M

jprocesses some job whose processing time is smaller than C

j!C

i, then

moving this job from Mjto M

idecreases +m

i/1C2

ibut does not increase the convex objective

function; another contradiction. Otherwise, Mjmust process at least two jobs with processing

times *Cj!C

i'C

i. Swapping one of these jobs with the set of jobs processed on M

iagain

decreases +mi/1

C2i

and does not increase the objective function; this is the final contradiction.j

2.1. The transformed instance

In this section, we introduce a so-called transformed instance that is a kind of rounded versionof the original instance I, and that can easily be solved in polynomial time. Indeed, let j be somefixed positive integer. Then the transformed instance Id(j) of scheduling instance I is defined asfollows:

(1) For every job Jjin I with p

j'¸/j, instance Id(j) contains a corresponding job Jd

jwhose

processing time pd

jequals p

jrounded up to the next integer multiple of ¸/j2. Note that for

the rounded processing time pj)pd

j)(j#1) p

j/j holds.

(2) Let S denote the total processing time of the jobs in I whose processing times are notgreater than ¸/j, and let Sd denote the value of S rounded up to the next integer multiple of¸/j. Then instance Id(j) contains Sdj/¸ new jobs, each of length ¸/j. Clearly, the totalnumber of jobs cannot increase and hence is bounded by n.

(3) The number of machines remains the same as in instance I.

Note that in instance Id(j), the processing times of all jobs are integer multiples of ¸/j2. Let¸

d"1/m + n

j/1pd

jdenote the average load of the machines in the transformed instance. By

construction,

¸)¸d. (3)

Since in I all jobs had lengths at most ¸ (and since ¸ is an integer multiple of ¸/j2), by theconstruction also all jobs in Id(j) have lengths smaller than ¸. From now on, jobs (in instanceI and in instance Id(j)) with processing times '¸/j will be called big jobs, and jobs withprocessing times )¸/j will be called small jobs.

Observation 2.3. In Id(j), no job has processing time greater than or equal to ¸d. ¹here exists an

optimum schedule in which all machine completion times fulfill 12¸

d(C

id(2¸d.

We now show two alternative methods to find an optimal solution, for instance, Id(j) inpolynomial time. The first method (cf. Theorem 2.4) is based on simple dynamic programmingand is fairly easy to implement. Its time complexity is a polynomial whose exponent depends on j.The second method (cf. Theorem 2.5) is a more efficient but also more complex algorithm which isbased on integer linear programming in a fixed dimension [17].

APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES 59

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 6: Approximation schemes for scheduling on parallel machines

First, observe that all jobs in Id(j) have processing times of the form k¸/j2 with j)k)j2. Weuse the following notation to represent the input and the assignments of jobs to machines: Theinput jobs are represented as a vector n"(nj, . . . , njÈ), where n

kdenotes the number of jobs

whose processing time equals k¸/j2. Notice that n*+jÈk/j

nk. An assignment to a machine is

a vector u"(uj, . . . , ujÈ), where uk

is the number of jobs of length k¸/j2 assigned to thatmachine. The length of assignment u, denoted C (u), is + jÈ

k/juk) k¸/j2. Denote by º the set of all

assignment vectors u with 12¸

d(C(u)(2¸d. It is easy to see that each assignment u3º consists

of at most 2j jobs and that therefore DºD)j4j holds. By Observation 2.3, when searching for anoptimum schedule we only need to consider assignment vectors in º.

First, we describe the dynamic programming approach. Let » be the set of vectors correspond-ing to subsets of the input, i.e. »"Mv : 0)v)nN. For every v3» and for every i, 0)i)m, wedenote by VAL(i, v) the total cost of the optimum assignment of the jobs in v to i machines thatonly uses assignments from º; in case no such assignment exists, VAL(i, v)"#R. It is easy to seethat VAL( ) , ) ) satisfies the following recurrence:

VAL(0, 0)"0

VAL(1, v)"Gf (C(v)) if v3º

#R if v Nº

and for i*2

VAL(i, v)" minv*u|U

M f (C (u))#VAL(i!1, v!u)N

For computing the value of a pair (i, v), we need to look at no more than DºD possible assignments,and for each such assignment we do a constant amount of work. Thus, the total running time ofthe resulting dynamic programming algorithm is O (m ) D» D ) DºD). Since D» D"<jÈ

k/1(n

k#1))njÈ,

we have proved the following theorem.

¹heorem 2.4. For any constant j*1, an optimal solution for the instance Id(j) of problemPD ) D+ f (C

i) can be computed in polynomial-time O(mnjÈ).

The second method to solve instance Id(j) is by using integer linear programming (ILP) witha fixed number of variables (cf. Reference 17). As before, º is the set of possible assignments toa machine. For each vector u3º, denote by xu the numbers of machines that were assigned u. Inthese terms, a feasible solution to the problem is a non-negative integer vector X such that+

u|Uxu"m, (i.e. m machines are used) and such that +

u|Uxu ) u"n (i.e. all jobs were assigned).

An optimum solution is given by the following integer system:

min +u|U

xu ) f (C(u))

s.t.

+u|U

xu"m

+u|U

xu ) u"n

xu*0 ∀u3º

60 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 7: Approximation schemes for scheduling on parallel machines

Notice that the dimension of this integer linear program is DºD)j4j and that the number ofconstraints is DºD#j2!j#2, i.e. two constants that do not depend on the input. The timecomplexity of Lenstra’s algorithm [17] is exponential in the dimension of the program butpolynomial in the logarithms of the coefficients. Therefore, its running time is O(logO(1)n) wherethe hidden constant depends exponentially on j. Since the program (ILP) can be constructed inO(n) time, we arrive at the following theorem.

¹heorem 2.5. For any constant j*1, an optimal solution for the instance Id(j) of problemPD ) D + f (C

i) can be computed in time O (n).

2.2. ¹ransformed instance versus original instance

In this section, we study the relationships between schedules for instance I and schedules forinstance Id(j).

¸emma 2.6. ¸et & be a schedule for instance I with machine completion times Ci, where

12¸(C

i(2¸. ¹hen there exists a schedule &d for instance Id(j) with machine completion times

Cd

i, where

Ci!

1

j¸)Cd

i)

j#1

jC

i#

1

j¸ (4)

Proof. In schedule & , replace every big job Jjby its corresponding job Jd

jin Id(j). This may

increase some machine completion times by a factor of at most (j#1)/j, but it never can decreasea machine completion time. Next, rearrange on every machine M

ithe jobs in such a way that the

small jobs are processed in the end, and let sidenote their total size. Clearly, +m

i/1si"S. There

exist integer multiples sd

iof ¸/j for 1)i)m, such that

Dsi!sd

iD)¸/j and

m+i/1

sd

i"Sd (5)

Removing the small jobs from the current schedule and putting sd

ij/¸ jobs of length ¸/j instead

of them on machine Miyields the desired schedule &d. j

¸emma 2.7. ¸et &d be a schedule, for instance, Id(j) with machine completion times Cd

i, where

12¸

d(C

id(2¸d. ¹hen there exists a schedule & , for instance, I with machine completion times

Ci, where

jj#1

Cd

i!

2

j¸)C

i)Cd

i#

1

j¸. (6)

Moreover, such a schedule & can be computed in O(n) time.

Proof. In schedule &d, replace every big job Jd

jby its corresponding job J

jin I. This cannot

increase the machine completion times; any machine completion time that is decreased isdecreased by at most a factor of j/(j#1). Next, let sd

idenote the number of jobs of length ¸/j

that are processed on machine Miin schedule &d; rearrange on every machine the jobs in such

a way that these small jobs are processed in the end. Below, we will show how to partition the

APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES 61

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 8: Approximation schemes for scheduling on parallel machines

small jobs in instance I into m parts S*1, . . . , S*

msuch that the total size of the jobs in part S*

iis

between (sd

i!2) ¸/j and (sd

i#1) ¸/j. Then replacing on every machine M

ithe sd

ijobs of length

¸/j by the small jobs in S*i

yields the desired schedule that fulfills inequality (6).In order to construct the partition S*

1, . . . , S*

m, one proceeds as follows. In the first partitioning

step, greedily assign to every set S*i

a number of small jobs from I until the total size of assignedjobs exceeds (sd

i!2) ¸/j. Note that the total size of jobs assigned to S*

iis bounded by (sd

i!1)

¸/j. Since

m+i/1

(sd

i!1) ¸/j)Sd

!m¸/j)S (7)

one can complete the first partitioning step without running out of jobs. In the second partition-ing step, greedily add the remaining small jobs to the sets S*

i, but without letting the total size

assigned to S*i

exceed (sd

i#1) ¸/j. It can be easily shown that indeed all small jobs are assigned

in the second partitioning step.

2.3. The approximation scheme

Finally, we will put everything together and derive the desired PTAS. Consider some fixed realnumber e, 0(e(1. Let d'0 be a real number that fulfills

f (y))A1#e3B f (x) for all x, y*0 with (1!d)x)y)(1#d)x (8)

Note that by condition (F*) such a d always exists. Define a positive integer j* :"v5/dw. Fora given instance I of P D ) D + f (C

i), the approximation scheme performs the following three Steps

(S-1)— (S-3).

(S-1) Compute the instance Id(j*) from I.(S-2) Solve instance Id(j*) to optimality; call the resulting schedule &d.(S-3) Transform schedule &d into a schedule & for the original instance I as described in the

proof of Lemma 2.7.

It is straightforward to perform Step (S-1) in linear time. By Theorem 2.5 and by Lemma 2.7, alsoSteps (S-2) and (S-3) can be implemented in O(n) time.

Now, let OPT denote the value of the objective function in the optimum schedule of I, letOPTd denote the value of the objective function for the optimum schedule of Id(j*), and letAPPROX denote the value of the objective function of the schedule for I that is computed in Step(S-3). First, we claim that

OPTd)A1#

e3B OPT (9)

holds: Indeed, by Observation 2.2, there exists an optimum schedule for I whose machinecompletion times C

ifulfill 1

2¸(C

i(2¸. Then by Lemma 2.6, there exists a schedule for Id(j*)

with machine completion times Cd

isuch that

DCi!Cd

iD)

1

j*C

i#

1

j*¸)

3

j*C

i)dC

i(10)

62 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 9: Approximation schemes for scheduling on parallel machines

By (8), f (Cd

i))(1#e/3) f (C

i). Summing up this inequality over all machines, one gets that the

objective value of the schedule for Id(j*) in Lemma 2.6 is at most (1#e/3) OPT. This implies thecorrectness of (9).

Next, we claim that

APPROX)A1#e3B OPT

d (11)

Indeed, by Observation 2.3, there exists an optimum schedule for Id(j*) whose machine comple-tion times Cd

ifulfill 1

d(C

id(2¸d. Then by Lemma 2.7, there exists a schedule for the

original instance I with machine completion times Cisuch that

DCi!Cd

iD)

1

j*#1Cd

i#

2

j*¸)

1

j*Cd

i#

2

j*¸

d)

5

j*Cd

i)dCd

i(12)

Similarly as above, inequality (8) now yields the correctness of (11). Finally, combining inequali-ties (9) and (11) yields that

APPROX)A1#e3B

2) OPT)(1#e) OPT (13)

Summarizing, the procedure (S-1)—(S-3) finds in polynomial time a schedule for I whose objectivevalue is at most a factor of (1#e) away from the optimum. This completes the proof ofTheorem 1.1 for problem P D ) D+ f (C

i).

3. DISCUSSION OF THE PTAS

In this section, we explain how to carry over the PTAS in Section 2 from problem (I) to problems(II), (I@) and (II@).

First, let us take a look at problem (II), i.e. minimizing maxmi/1

f (Ci). It is easy to see that an

analogous statement as in Observation 2.2 still holds true for problem (II), i.e. there is anoptimum schedule in which all C

iare between 1

2¸ and 2¸. The transformed instance is formulated

as in Section 2.1 and solved in O(n) time similarly as in Theorem 2.5. The only difference is thatone has to use a slightly different objective function in the integer linear program (ILP):

min zs. t.

+u|U

xu"m

+u|U

xu ) u"n

xu)m ) yu

z*yu ) f (C(u))

0)xu , xu integer ∀u3º

yu3M0, 1N ∀ u3º

APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES 63

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 10: Approximation schemes for scheduling on parallel machines

The 0—1 variables yu indicate whether the assignment u is used (yu"1) or not (yu"0). In case theassignment u is used, then the term yu ) f (C (u)) contributes to the objective function. Finally, thearguments in Section 2.3 can be used to show that Steps (S-1)— (S-3) yield a (1#e) -approximationalgorithm.

For the maximization problems (I@) and (II@), again a statement as in Observation 2.2 holds true,i.e. there always is an optimum schedule in which all C

iare between 1

2¸ and 2¸ (recall that now

we are dealing with concave functions). The transformed instance is formulated as in Section 2.1.It can be solved (with an appropriately modified objective function) in O(n) time via Lenstra’salgorithm as in Theorem 2.5. Similarly as in Section 2.3, one can argue that Steps (S-1)—(S-3)always yield a schedule that is at most a factor of (1!e) away from the optimum. The details areleft to the reader. This completes the proof of Theorem 1.2.

Finally, let us show that a straightforward application of the methods of Sahni [4] and ofHochbaum and Shmoys [5] cannot yield a proof of Theorems 1.1 and 1.2. This method in [4, 5]first removes the small jobs that are smaller than some threshold q, then solves the remainingproblem to get an intermediate schedule, and at last greedily adds the small jobs to thisintermediate schedule.

Example 3.1. Consider an instance with m"3 machines, 6 jobs with lengths 13, 9, 9, 6, 6, 6and 10/q tiny jobs, each of length q/2; note that the overall length of the tiny jobs is 5. Theobjective is to minimize + C2

i. It is straightforward to verify that the optimum schedule assigns to

every machine a total processing time of 18: three jobs of length 6 to machine M1, two jobs of

length 9 to machine M2, and the remaining jobs to machine M

3. The optimum objective value

equals 972. If one removes the tiny jobs, then the unique optimum schedule for the jobs13, 9, 9, 6, 6, 6 is to assign 13 and 6 to machine M

1, 9 and 6 to M

2, and 9 and 6 to M

3. No matter,

how one adds the tiny jobs to this intermediate schedule, one can never reach a schedule withobjective value better than 9731

2.

4. APPROXIMATION SCHEMES, CONDITION (F*), AND THE LPT

In this section, we investigate the scheduling problem (I), i.e. problem P D ) D+ f (Ci). We discuss the

impact of property (F*) for functions on the approximability behaviour of problem P D ) D + f (Ci).

Moreover, we discuss the relationship between the existence of a PTAS and the worst-casebehaviour of the LPT rule.

First, let us consider the function f (x)"2x, x3R`0. It can easily be seen that function 2x is

non-negative and convex, but does not satisfy property (F*). Hence, the following result demon-strates that our main result in Theorem 1.1 does not hold true without imposing condition (F*).

Observation 4.1. ºnless P"NP, the problem P D ) D + 2Ci does not allow a polynomial-timeapproximation algorithm with finite performance guarantee. Especially, the problem does not allowa P¹AS.

Proof. Suppose that there exists a polynomial time approximation algorithm with finiteperformance guarantee o'1. Define an integer a"vlog

2ow#2. For an integer Z and for a set

z1, . . . , z

nof positive integers with + n

j/1zj"2Z, it is NP-complete to decide whether the z

jcan

be partitioned into two sets whose elements sum up to Z (see [16]).Define a corresponding scheduling instance with m"2 machines and n jobs J

j, 1)j)n, with

lengths pj"az

j. It is easy to see that exactly one of the following two cases holds: (i) The z

jcan be

64 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 11: Approximation schemes for scheduling on parallel machines

partitioned into two sets whose elements sum up to Z and, hence, the optimum objective valueequals 2aZ`1. (ii) The z

jcan not be partitioned into two sets whose elements sum up to Z and,

hence, the optimum objective value is at least 2aZ`a. Since the approximation algorithmcould distinguish between these two cases in polynomial time, it would be able to solve theNP-complete partition problem in polynomial time. The claim follows.

Note that the non-approximability result in Observation 4.1 does not only rely on theexponential growth rate of function 2x : An analogous statement can be proved e.g. for functionf (x)"(x!1)2 . We do not know whether Theorem 1.1 still holds true without imposingconvexity of function f .

Next, let us discuss the relationship between the existence of an approximation scheme fora scheduling problem, and between the behavior of the LPT rule (cf. Section 1.1) for this problem.We offer the following conjecture.

Conjecture 4.2. ¸et f be a non-negative, convex function R`0PR`

0and let P D ) D + f (C

i) be the

corresponding scheduling problem. ¹hen the following three statement are all equivalent.

(i) P D ) D + f (Ci) possesses a polynomial time approximation algorithm with finite performance

guarantee.(ii) For P D ) D + f (C

i), the ¸P¹ rule has finite performance guarantee.

(iii) P D ) D + f (Ci) possesses a P¹AS.

We are able to prove this conjecture for the case where f is not only convex but also satisfiesproperty (F*). In this case, Theorem 1.1 implies the truth of (iii) and (i), and the followingobservation yields the truth of (ii).

Observation 4.3. ¸et f be a non-negative, convex function R`0PR`

0that satisfies property (F*).

¹hen the ¸P¹ rule for problem P D ) D + f (Ci) has finite performance guarantee.

Proof. Let ¸"(1/m)+ nj/1

pj. Since the LPT rule assigns every job with processing *¸ to its

own machine, we may assume by arguments that are similar to those applied at the beginning ofSection 2 that no job has processing time greater than or equal to ¸. Next, it can be shown that inthe schedule produced by LPT, every machine completion time C

ifulfills 1

2¸(C

i(2¸. By

Observation 2.2, an analogous inequality holds for the machine completion times in an optimumschedule. Consequently, every machine completion time in the LPT schedule is at most a factor 4away from the corresponding machine completion time in the optimum schedule.

Now, consider condition (F*) with e"1: We get that there exists a d'0, such that for ally)(1#d)x the inequality f (y))2 f (x) holds. Define k as the smallest integer that fulfills(1#d)k*4. Then for all y)4x the inequality f (y))2k f (x) holds. Combining this inequalitywith the above discussion on the LPT schedule, one gets that the LPT algorithm indeed has finiteperformance guarantee 2k . j

For general functions f, the worst-case behaviour of the LPT rule does not seem to be a goodindicator for the approximability of a scheduling problem: It is easy to find non-convex functionsf for which problem P D ) D + f (C

i) possesses a PTAS or is even solvable in polynomial time, whereas

the LPT rule does not have finite performance guarantee; take e.g. f (x)"ln(1#x). Moreover,there exist non-convex functions f for which problem P D ) D + f (C

i) does not possesse a PTAS

(unless P"NP) whereas the LPT rule does have finite performance guarantee; take e.g. thefunction f with f (1)"1 and f (x)"2 for xO1.

APPROXIMATION SCHEMES FOR SCHEDULING ON PARALLEL MACHINES 65

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)

Page 12: Approximation schemes for scheduling on parallel machines

5. CONCLUSION

We have shown that for a large set of scheduling problems on parallel machines whose objectivefunctions solely depend on the machine completion times, there exist polynomial-time approxi-mation schemes. The running times of our algorithms are linear in the number of jobs andmachines. Our result covers as special cases the approximation schemes of Sahni [4], Hochbaumand Shmoys [5], Alon et al. [10], and Woeginger [15].

We pose as an open problem to derive similar results for scheduling problems on parallelmachines whose objective functions solely depend on the job completion times.

REFERENCES

1. E. L. Lawler, J. K. Lenstra, A. H. G. Rinnooy Kan and D. B. Shmoys, ‘Sequencing and scheduling: algorithms andcomplexity’, in Handbooks in Operations Research and Management Science, Vol. 4, North Holland, Amsterdam,1993, pp. 445—522.

2. R. L. Graham, ‘Bounds for certain multiprocessor anomalies’, Bell System ¹ech. J., 45, 1563—1581 (1966).3. R. L. Graham, ‘Bounds on multiprocessing timing anomalies’, SIAM J. Appl. Math., 17, 416—429 (1969).4. S. Sahni, ‘Algorithms for scheduling independent tasks’, J. Assoc. Comput. Mach., 23, 116—127 (1976).5. D. S. Hochbaum and D. B. Shmoys, ‘Using dual approximation algorithms for scheduling problems: theoretical and

practical results’, J. Assoc. Comput. Mach., 34, 144—162 (1987).6. R. A. Cody and E. G. Coffman, ‘Record allocation for minimizing expected retrieval costs on drum-like storage

devices’, J. Assoc. Comput. Mach., 23, 103—115 (1976).7. A. K. Chandra and C. K. Wong, ‘Worst-case analysis of a placement algorithm related to storage allocation’, SIAM J.

Comput., 4, 249—263 (1975).8. J. Y. T. Leung and W. D. Wei, ‘Tighter bounds on a heuristic for a partition problem’, Inform. Process.¸ett., 56, 51—57

(1995).9. A. Avidor, Y. Azar and J. Sgall, ‘Ancient and new algorithms for load balancing in the l

pnorm’, in Proc. 9th Symp. on

Discrete Algorithms, 1998, pp. 426—435.10. N. Alon, Y. Azar, G. J. Woeginger and T. Yadid, ‘Approximation schemes for scheduling’, in Proc. 8th Symp. on

Discrete Algorithms, 1997, pp. 493—500.11. P. Dell’Olmo, H. Kellerer, M. G. Speranza and Zs. Tuza, ‘Partitioning items in a fixed number of bins to minimize the

total size’, Manuscript, University of Brescia, Italy, 1996.12. D. K. Friesen and B. L. Deuermeyer, ‘Analysis of greedy solutions for a replacement part sequencing problem’, Math.

Oper. Res., 6, 74—87 (1981).13. B. L. Deuermeyer, D. K. Friesen and M. A. Langston, ‘Scheduling to maximize the minimum processor finish time in

a multiprocessor system’, SIAM J. Alg. Discrete Meth., 3, 190—196 (1982).14. J. Csirik, H. Kellerer and G. J. Woeginger, ‘The exact LPT-bound for maximizing the minimum completion time’,

Oper. Res. ¸ett., 11, 281—287 (1992).15. G. J. Woeginger, ‘A polynomial time approximation scheme for maximizing the minimum machine completion time’,

Oper. Res. ¸ett., 20, 149—154 (1997).16. M. R. Garey and D. S. Johnson, Computers and Intractability, Freeman and Company, San Francisco, 1979.17. H. W. Lenstra Jr., ‘Integer programming with a fixed number of variables’, Math. Oper. Res., 8, 538—548 (1983).

66 N. ALON E¹ A¸.

( 1998 John Wiley & Sons, Ltd. J. Sched. 1, 55—66 (1998)