Abstract
We study kill-and-restart and preemptive strategies for the fundamental scheduling problem of minimizing the sum of weighted completion times on a single machine in the non-clairvoyant setting. First, we show a lower bound of 3 for any deterministic non-clairvoyant kill-and-restart strategy. Then, we give for any \(b > 1\) a tight analysis for the natural b-scaling kill-and-restart strategy as well as for a randomized variant of it. In particular, we show a competitive ratio of \((1+3\sqrt{3})\approx 6.197\) for the deterministic and of \(\approx 3.032\) for the randomized strategy, by making use of the largest eigenvalue of a Toeplitz matrix. In addition, we show that the preemptive Weighted Shortest Elapsed Time First (WSETF) rule is 2-competitive when jobs are released online, matching the lower bound for the unit weight case with trivial release dates for any non-clairvoyant algorithm. Using this result as well as the competitiveness of round-robin for multiple machines, we prove performance guarantees smaller than 10 for adaptions of the b-scaling strategy to online release dates and unweighted jobs on identical parallel machines.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Minimizing the total weighted completion time on a single processor is one of the most fundamental problems in the field of machine scheduling. The input consists of n jobs with processing times \(p_1,\ldots ,p_n\) and weights \(w_1,\dotsc ,w_n\), and the task is to sequence them in such a way that the sum of weighted completion times \(\sum _{j=1}^n w_j C_j\) is minimized. We denote this problem as . Smith [2] showed in the 50’s that the optimal schedule is obtained by the Weighted Shortest Processing Time first (\(\text {WSPT}\)) rule, i.e., jobs are sequenced in non-decreasing order of the ratio of their processing time and their weight.
Reality does not always provide all information beforehand. Around 30 years ago, the non-clairvoyant model, in which the processing time of any job becomes known only upon its completion, was introduced for several scheduling problems [3,4,5]. It is easy to see that no non-preemptive non-clairvoyant algorithm can be constant-competitive for the unweighted variant . In their seminal work, Motwani et al. [5] proved for this problem that allowing preemption breaks the non-constant barrier. Specifically, they showed that the natural round-robin algorithm is 2-competitive, matching a lower bound for all non-clairvoyant algorithms. This opened up a new research direction, leading to constant-competitive preemptive non-clairvoyant algorithms in much more general settings, like weighted jobs [6], multiple machines [7,8,9], precedence constraints [10, 11], and non-trivial release dates. When jobs are released over time, they are assumed to be unknown before their arrivals (online scheduling). No lower bound better than 2 is known for this case, whereas the best known upper bound before this work was 3, see e.g. [12].
But there is a downside of the preemptive paradigm as it uses an unlimited number of interruptions at no cost and has a huge memory requirement to maintain the ability to resume all interrupted jobs. Therefore, we continue by studying the natural class of kill-and-restart strategies that—inspired by computer processes—can abort the execution of a job (kill), but when processed again later, the job has to be re-executed from the beginning (restart). It can be considered as an intermediate category of algorithms between preemptive and non-preemptive ones, as on one hand jobs may be interrupted, and on the other hand when jobs are completed, they have been processed as a whole. Hence, by removing all aborted executions one obtains a non-preemptive schedule. Although this class of algorithms has already been investigated since the 90’s [4], to the best of our knowledge, the competitive ratio of non-clairvoyant kill-and-restart strategies for the total completion time objective has never been studied.
Our contribution We start by strengthening the preemptive lower bound of 2 for the kill-and-restart model.
Theorem 2
For , no deterministic non-clairvoyant kill-and-restart strategy can achieve a competitive ratio smaller than \(3-\frac{2}{n+1}\) on instances with \(n\ge 3\) jobs, even if every job j has processing time \(p_j \ge 1\).
The main part of this work is devoted to the b-scaling strategy \({\mathfrak {D}_{b}}\) that repeatedly probes each unfinished job for the time of an integer power of \(b>1\) multiplied by its weight. For it is easy to see that \(\mathfrak {D}_{2}\) is 8-competitive by comparing its schedule to the weighted round-robin schedule for a modified instance and using the 2-competitiveness due to Kim and Chwa [6]. Using a novel and involved analysis we determine the exact competitive ratio of \({\mathfrak {D}_{b}}\).
Theorem 6
For \(b>1\), \({\mathfrak {D}_{b}}\) is \(\bigl ( 1+\frac{2b^{\nicefrac {3}{2}}}{b-1} \bigr )\)-competitive for . This ratio is minimized for \(b=3\), yielding a performance guarantee of \(1+3\sqrt{3}\approx 6.196\).
Theorem 7
For every \(b > 1\), there exists a sequence of instances \((\varvec{p}_L)_{L \in \mathbb {N}}\) for such that

Our main technique is to reduce the problem of finding the competitive ratio of \({\mathfrak {D}_{b}}\) to the computation of the largest eigenvalue of a tridiagonal Toeplitz matrix. Subsequently, we obtain a significantly better exact competitive ratio for a randomized version of the b-scaling strategy, denoted by \({\mathfrak {R}_b}\), that permutes the jobs uniformly at random and chooses a random offset drawn from a log-uniform distribution.
Theorem 10
For every \(b>1\), \({\mathfrak {R}_b}\) is \(\frac{2b+\sqrt{b}-1}{\sqrt{b} \ln b}\)-competitive for . This ratio is minimized for \(b\approx 8.16\), yielding a performance guarantee smaller than 3.032.
Theorem 11
For all \(b>1\), there exists a sequence of instances \((\varvec{p}_L)_{L\in \mathbb {N}}\) for such that
The analysis basically mimics that of the deterministic strategy, but it is necessary to group the jobs whose Smith ratio falls in the ith interval of the form \((b^{i/K},b^{(i+1)/K}]\), where K is a large natural number. This approach leads to the computation of the largest eigenvalue of a banded symmetric Toeplitz matrix of bandwidth \(2K-1\), and the result is obtained by letting \(K\rightarrow \infty \).
We then study more general scheduling environments. For the online problem, in which jobs are released over time, denoted by , we close the gap for the best competitive ratio of preemptive algorithms by analyzing the Weighted Shortest Elapsed Time First rule, short \(\text {WSETF}\). This policy runs at every point in time the job(s) with minimum ratio of the processing time experienced so far (elapsed time) over the weight.
Theorem 12
\(\text {WSETF}\) is 2-competitive for .
Theorem 12 generalizes the known 2-competitiveness for trivial release dates shown by Kim and Chwa [6]. It also matches the performance guarantee of the best known stochastic online scheduling policy \(\text {F-GIPP}\) [13], a generalization of the Gittins index priority policy [14, 15], for the stochastic variant of our problem where the probability distributions of the processing times are given at the release dates and the expected objective value is to be minimized. Our improvement upon the analysis of this policy, applied to a single machine, is threefold: First, our strategy does not require any information about the distributions of the processing times, second, we compare to the clairvoyant optimum, while \(\text {F-GIPP}\) is compared to the optimal non-anticipatory policy, and third, \(\text {WSETF}\) is more intuitive and easier to implement in applications than the \(\text {F-GIPP}\) policy.
Using Theorem 12, we then give an upper bound on the competitive ratio of a generalized version of \({\mathfrak {D}_{b}}\) that also handles jobs arriving over time by never interrupting a probe.
Theorem 19
\({\mathfrak {D}_{b}}\) is \(\frac{2b^4}{2b^2-3b+1}\)-competitive for . This ratio is minimized for \(b=\frac{9+\sqrt{17}}{8}\), yielding a performance guarantee of \(\frac{107+51\sqrt{17}}{32} \approx 9.915\).
Finally, we also analyze the unweighted problem on multiple identical parallel machines.
Theorem 23
\({\mathfrak {D}_{b}}\) is \(\frac{3b^2-b}{b-1}\)-competitive for . This ratio is minimized for \(b=\frac{3+\sqrt{6}}{3}\), yielding a performance guarantee of \(5+2\sqrt{6} \approx 9.899\).
Related work
Non-preemptive scheduling. The beginnings of the field of machine scheduling date back to the work of Smith [2], who investigated the problem of non-preemptively minimizing the sum of weighted completion times on a single machine. Its optimal schedule is obtained by sequencing the jobs in non-decreasing order of their processing time to weight ratio \(\nicefrac {p_j}{w_j}\) (Smith’s rule). When all jobs have unit weights, one obtains the Shortest Processing Time first (SPT) rule. This can be generalized to the identical parallel machine setting, where list scheduling [16] according to SPT is optimal [17] for unit-weight jobs. However, the problem of scheduling jobs released over time on a single machine is strongly NP-hard [18] (even for unit weights), and Chekuri and Khanna developed a polynomial-time approximation scheme (PTAS) for it [19]. When jobs arrive online, no deterministic algorithm can be better than 2-competitive [20], and this ratio is achieved by a delayed variant of Smith’s rule [21].
In the non-clairvoyant setting it is well known that no (randomized) non-preemptive algorithm is constant-competitive (see Proposition 1). A less pessimistic model is the stochastic model, where the distributions of the random processing times \(P_j\) are given and one is interested in non-anticipatory (measurable) policies [22]. For some classes of distributions, this information allows obtaining constant expected competitive ratios [23] for parallel identical machines. However, policies minimizing the expected competitive ratio do not need to minimize the expected objective value—the classic measure in stochastic optimization. For this criterion, Rothkopf [24] showed that for the single machine case the optimality of Smith’s rule can be transferred to the Weighted Shortest Expected Processing Time rule, in which jobs are sorted in non-decreasing order of \(\nicefrac {\mathbb {E}[P_j]}{w_j}\). In order to deal with the stochastic counterparts of the NP-hard problems mentioned above, Möhring et al. [25] introduced approximative scheduling policies, whose expected objective value is compared to the expected objective value of an optimal non-anticipatory policy. While there are constant-competitive policies for stochastic online scheduling on a single machine [26], the performance guarantees of all known approximative policies for multiple machines depend on either the maximum coefficient of variation [27] or the number of jobs and machines [28], even for unit-weight jobs released at time 0.
Preemptive scheduling. For the clairvoyant offline model, allowing preemption only helps in the presence of non-trivial release dates [29]. In this case, the optimal preemptive schedule may be a factor of \(\text {e}/(\text {e} - 1)\) better than the best non-preemptive one [30]. Finding an optimal preemptive schedule is still strongly NP-hard [31], and there is a PTAS adapted to this problem [19]. For jobs arriving online Sitters [32] developed a 1.566-competitive deterministic algorithm, and Epstein and van Stee [33] proved a lower bound of 1.073.
When the job lengths are uncertain, allowing preemption becomes much more crucial. Motwani et al. [5] showed that the simple (non-clairvoyant) round-robin procedure has a competitive ratio of 2 for minimizing the total completion time on identical machines. This gives the same share of machine time to each job in rounds of infinitesimally small time slices. For weighted jobs, the Weighted Round-Robin (\(\text {WRR}\)) rule (also known as generalized processor sharing (GPS) [34, 35]), which always distributes the available machine capacity to the jobs proportionally to their weights, was shown to be 2-competitive on a single machine by Kim and Chwa [6], and the same competitive ratio is achieved by a generalization for multiple identical machines [7]. Similar time sharing algorithms were also developed in the context of non-clairvoyant online scheduling, where jobs arrive over time and are not known before their release dates. Here one can distinguish between minimizing the total (weighted) completion time and the total (weighted) flow time. The \(\text {WRR}\) rule can be generalized in two natural way in this setting: Either the machine capacity is still allocated based only on the weights or based on the weighted elapsed times, resulting in the \(\text {WSETF}\) rule, mentioned above. It is easy to see that both are 3-competitive, see e.g. [12]. On the other hand, there exist examples showing that the first option is not 2-competitive for total weighted completion time. For the total weighted flow time objective constant competitiveness is unattainable [5]. Apart from work on non-constant competitive ratios [36], the problem has been primarily studied in the resource augmentation model [37], where the machine used by the algorithm runs \(1+\varepsilon \) times faster. Kim and Chwa and Bansal and Dhamdhere [38] independently proved that \(\text {WSETF}\) is \((1+\varepsilon )\)-speed \((1+\nicefrac 1 \varepsilon )\)-competitive for weighted flow time on a single machine. By running this algorithm on the original-speed machine, the completion times increase by a factor of \(1+\nicefrac 1 \varepsilon \), so that one obtains a \((1+\varepsilon )(1+\nicefrac 1 \varepsilon )\)-competitive algorithm for the total weighted completion time [39], which yields a ratio of 4 for \(\varepsilon = 1\). The proofs of Kim and Chwa and Bansal and Dhamdhere both proceed by showing that at any time \(t \ge 0\) the total weight of unfinished jobs in the \(\text {WSETF}\) schedule is at most a factor of \((1+\nicefrac 1 \varepsilon )\) larger than the unfinished weight in the optimal schedule. The lower-bound example of Motwani et al. (many equal small jobs released at time 0) demonstrates that with such an approach no better bound than 4 is achievable. Consequently, a completely different technique is needed to prove Theorem 12. For the much more general setting of unrelated machines Im et al. [9] established a \((1+\varepsilon )\)-speed \(\mathcal {O}(\nicefrac {1}{\varepsilon ^2})\)-competitive algorithm. Motwani et al. also considered the model in which the number of allowed preemptions is limited, for which they devised algorithms that resemble the kill-and-restart algorithms presented in this paper. As mentioned above, for the stochastic model for minimizing the expected total weighted completion time, the Gittins index policy is optimal for single-machine with trivial release dates [14, 15], and Megow and Vredeveld [13] established a 2-competitive online policy for multiple machines and arbitrary release dates.
Kill-and-restart scheduling. The kill-and-restart model was introduced by Shmoys et al. [4] in the context of makespan minimization. For the total completion time objective we are not aware of any work on kill-and-restart strategies in the non-clairvoyant model. However, in the clairvoyant online model, kill-and-restart algorithms have been considered by Vestjens [40] and Epstein and van Stee [33], who gave lower bounds that are larger than the lower bounds for preemptive algorithms but much smaller than the known lower bounds for non-preemptive online algorithms, suggesting that allowing restarts may help in the online model. The proof of this fact was given several years later by van Stee and La Poutré [41], who achieved a deterministic competitive ratio of 3/2 for minimizing the total completion time on a single machine, beating even the lower bound of \(\text {e}/(\text {e} - 1) \approx 1.582\) for any randomized non-preemptive online algorithm [42]. In the non-clairvoyant setting, considered in this work, we observe a much larger benefit from allowing restarts, reducing the competitive ratio from \(\Omega (n)\) to a constant.
Further related work. In the end, all aborted probes served only the purpose of collecting information about the unknown processing times of the jobs. Kill-and-restart strategies can thus be regarded as online algorithms for non-preemptive scheduling with the possibility to invest time in order to obtain some information. In that sense, the considered model resembles that of explorable uncertainty [43,44,45]. In order to allow for any reasonable competitiveness results, it must be ensured in both models that testing/probing provides some benefit to the algorithm other than information gain. In the explorable uncertainty model, this is achieved by the assumption that testing can shorten the actual processing times, while in our model the probing time replaces the processing time if the probe was long enough.
Scheduling on a single machine under the kill-and-restart model shares many similarities with optimal search problems, in which a number of agents are placed in some environment and must either find some target or meet each other as quickly as possible. A problem that received a lot of attention is the so-called w-lanes cow-path problem, in which an agent (the cow) is initially placed at the crossing of w roads, and must find a goal (a grazing field) located at some unknown distance on one the w roads [46]. For the case \(w=2\), deterministic and randomized search strategies were given that achieve the optimal competitive ratio of 9 [47] and approximately 4.5911 [47, 48], respectively. This work has been extended by Kao et al. [49], who give optimal deterministic and randomized algorithms for all \(w\in \mathbb {N}\). The single-machine scheduling problem with kill-and-restart strategies can in fact be viewed in this framework: There are now \(n=w\) goals, and the jth goal is located at some unknown distance \(p_j\) on the jth road. The agent can move at unit speed on any of the roads, and has the ability to teleport back to the origin at any point in time, which represents the action of aborting a job. The objective is to minimize the sum of times at which each goal is found.
2 Preliminaries
We consider the machine scheduling problem of minimizing the weighted sum of completion times on a single machine (). Formally, we consider instances \(I = (\varvec{p},\varvec{w})\) consisting of a vector of processing times \(\varvec{p} = (p_j)_{j=1}^n\) and a vector of weights \(\varvec{w} = (w_j)_{j=1}^n\).
If the jobs are in WSPT order, i.e., jobs are ordered increasingly by their Smith ratios \(p_j/w_j\), then it is easy to see that sequencing the jobs in this ordering yields an optimal schedule. We denote this (clairvoyant) schedule by \(\text {OPT}(I)\). By slight abuse of notation, we also denote the objective value of an optimal schedule by \(\text {OPT}(I)\). In particular, its cost is \( \text {OPT}(I) = \sum _{j=1}^n w_j\sum _{k=1}^j p_{k} = \sum _{j=1}^n p_{j}\sum _{k=j}^n w_j \).
The focus of our work lies in the analysis of non-clairvoyant strategies. We call a strategy non-clairvoyant if it does not use information on the processing time \(p_j\) of a job j before j has been completed. A deterministic strategy \(\mathfrak {D}\) is said to be c-competitive if, for all instances \(I = (\varvec{p}, \varvec{w})\), \(\mathfrak {D} (I) \le c \cdot \text {OPT}(I)\), where \(\mathfrak {D} (I)\) denotes the cost of the strategy for instance I. The competitive ratio of \(\mathfrak {D}\) is defined as the infimum over all c such that \(\mathfrak {D}\) is c-competitive. For a randomized strategy \(\mathfrak {R}\), the cost for instance I is a random variable \(X_I:\Omega \rightarrow \mathbb {R}_{\ge 0}\) that associates an outcome \(\omega \) of the strategy’s sample space to the realized cost, and we denote by \(\mathfrak {R}(I){:}{=}\mathbb {E}[X_I]\) the expected cost of the randomized strategy for instance I. We say that \(\mathfrak {R}\) is c-competitive if for all instances \(I = (\varvec{p}, \varvec{w})\), \(\mathfrak {R}(I)\le c \cdot \text {OPT}(I)\). It is well known that for our problem no non-preemptive strategy can achieve a constant competitive ratio.
Proposition 1
No randomized non-preemptive non-clairvoyant strategy has a constant competitive ratio for .
Proof
By Yao’s principle [50] it suffices to construct a randomized instance for which any deterministic strategy has non-constant competitive ratio. To this end, we consider the instance with n jobs where \(p_1=\cdots =p_{n-1}=1\) and \(p_n=n^2\) and randomize uniformly over all permutations of the jobs. Clearly, an optimal clairvoyant strategy sequences the jobs in any realization in SPT order and hence, we have \( \text {OPT}=\sum _{j=1}^n (n-j+1)p_j=\frac{1}{2}n(n-1)+n-1+n^2=\mathcal {O}(n^2). \)
The schedule of any deterministic strategy can be represented as a permutation of the jobs as idling only increases the objective value. Hence, for any permutation we obtain the expected cost
where we used the fact that in a uniformly distributed permutation, the probability that the long job appears in each position \(k\in [n]\) is \(\frac{1}{n}\). \(\square \)
Kill-and-restart strategies Due to this negative result, we study non-clairvoyant kill-and-restart strategies for that may abort the processing of a job, but when it is processed again later, it has to be executed from the beginning. In order to define such strategies, we first introduce a state and action space as well as a transition function modeling the kill-and-restart setting. Then, we can describe kill-and-restart strategies as functions mapping states to actions.
Formally, we consider the state space \(\mathcal {S}{:}{=}\mathbb {R}\times 2^{[n]} \times \mathbb {R}^n\). A state \((\theta , U, \varvec{\mu }) \in \mathcal {S}\) consists of the current time \(\theta \), the set of unfinished jobs U at \(\theta \), and a vector \(\varvec{\mu }\) of lower bounds on the processing times learned from past probes, such that \(p_j\ge \mu _j\) for all jobs j. For every state \(s = (\theta , U, \varvec{\mu }) \in \mathcal {S}\), there is a set of possible kill-and-restart actions \(\mathcal {A}(s)\), where an action \(a = \bigl ((t_i,j_i,\tau _i)\bigr )_{i \in \mathcal {I}} \in \mathcal {A}(s)\) is a family of probes \((t_i,j_i,\tau _i)\) such that the intervals \( (t_i, t_i+\tau _i)\), \(i \in \mathcal {I}\), are disjoint and contained in \(\mathbb {R}_{> \theta }\) and \(j_i \in U\) for all \(i \in \mathcal I\). We denote by \(\mathcal {A}= \bigcup _{s \in \mathcal {S}} \mathcal {A}(s)\) the set of all actions in all states. Additionally, we define a transition function \(T_I :\mathcal {S}\times \mathcal {A}\rightarrow \mathcal {S}\) depending on the instance I. This function transforms any state \(s = (\theta , U, \varvec{\mu })\) and action \(a = \big ((t_i, j_i, \tau _i)\big )_{i \in \mathcal {I}} \in \mathcal A(s)\) into a new state \(s' = (\theta ', U', \varvec{\mu }')\) as follows. First, we identify the probe indexed with \(i^* {:}{=}{\text {argmin}} \big \{ t_i + p_{j_i} \mid i \in \mathcal {I} \text { with } \tau _i \ge p_{j_i} \big \}\), which is the first probe in a that leads to the completion of some job. Then, the lower bounds \(\varvec{\mu }'\) of the new state \(s'\) are defined by \(\mu _j' {:}{=}\max \big \{ \mu _j, \max \{ \min \{ \tau _i, p_j \} \mid i \in \mathcal {I}: t_i \le t_{i^*}, j_i = j \} \big \}\), i.e., the lower bounds are updated to the new probing times or the processing times if appropriate. Further, the job completing in probe \(i^*\) is removed from the set of unfinished jobs by setting \(U'{:}{=}U{\setminus } \{j_{i^*}\}\), and the time is updated to \(\theta ' {:}{=}t_{i^*} + p_{j_{i^*}}\). Finally, we define a kill-and-restart strategy as a function \(\Pi :\mathcal {S}\rightarrow \mathcal {A}\) with \(\Pi (s) \in \mathcal {A}(s)\) for all \(s \in \mathcal {S}\). Note that a kill-and-restart strategy is non-clairvoyant by definition as it only has access to the lower bounds on the processing times, while the actual processing time is only revealed to the strategy upon completion of a job.
However, observe that such strategies may not be implementable, e.g., on a Turing machine, as the above definition allows for an infinite number of probes in a bounded time range. On the other hand, a deterministic kill-and-restart strategy without infinitesimal probing cannot be constant-competitive. To see this, consider an arbitrary algorithm ALG, and assume without loss of generality that the first job it probes is the first job presented in the input. Denote by \(t>0\) the first probing time, and, for \(\varepsilon < 1\), consider the instance \(I_\varepsilon =\big ((t, \varepsilon t,\ldots ,\varepsilon t),(1,\ldots ,1)\big )\) with n unit weight jobs. By construction, ALG processes the first job without aborting it, so \(ALG\ge nt + n(n-1)/2\cdot t \varepsilon = nt+\mathcal {O}(\varepsilon )\). On the other hand \(\text {OPT}\) schedules the job in SPT order, yielding \(\text {OPT}=n(n-1)/2\cdot t \varepsilon + (n-1)t\varepsilon +t=t+\mathcal {O}(\varepsilon )\). Hence, \(ALG/\text {OPT}\) approaches n as \(\varepsilon \rightarrow 0\). This subtlety is in fact inherent to all scheduling problems with unknown processing times or search problems with unknown distances. Of course, this can be avoided if a lower bound on the processing times \(p_j\) is given, e.g. \(p_j \ge 1\) for all \(j \in [n]\). In this case, the strategies analyzed in this paper can be turned into implementable ones. Throughout this paper, however, we don’t want to make this assumption and hence analyze strategies with infinitesimal probing. This way of describing strategies for search problems is commonly used in the literature (see, e.g. [47]) and simplifies the exposition of our analysis since some constants resulting from finite geometric series vanish when infinitesimal probing is introduced so that the geometric series are infinite.
We denote by \(Y_j^{\mathfrak {S}}(I, t)\) the total time for which the machine has been busy processing job j until time t in the schedule constructed by the strategy \(\mathfrak {S}\) on the instance I.
3 Lower bound for deterministic strategies
Theorem 2
For , no deterministic non-clairvoyant kill-and-restart strategy can achieve a competitive ratio smaller than \(3-\frac{2}{n+1}\) on instances with \(n\ge 3\) jobs, even if every job j has processing time \(p_j \ge 1\).
Proof
Let \(\varepsilon \in \bigl (\frac{2}{n+1},1\bigr ]\) and define \(T {:}{=}\frac{(2-\varepsilon )(n^2+n)}{\varepsilon (n+1)-2}\). Consider an arbitrary deterministic kill-and-restart strategy \(\mathfrak {D}\) with the initially chosen family of probes \((t_i,j_i,\tau _i)_{i\in \mathcal {I}}\). Let \(Y_j(\theta ){:}{=}\sum _{i\in \mathcal {I}: t_i< \theta , j_{i}=j } \min \{\tau _i,\theta -t_i\}\) be the total probing time assigned by \(\mathfrak {D}\) to job j up to time \(\theta .\) We define an instance \(I{:}{=}(\varvec{p},\varvec{1})\) by distinguishing two cases on the first job \(j_0\) planned to be probed at or after time T. Note that such a job exists, as otherwise \(\mathfrak {D}\) does not complete all jobs if processing times are long enough.
If \(j_0\) is probed for a finite amount of time, we denote by \(t\ge T\) the end of its probing time. Then, define \(p_j:=1+Y_j(t)\) for all \(j\in [n]\). Clearly, no job finishes before t when \(\mathfrak {D}\) runs the instance I, hence \(\mathfrak {D}(I) \ge nt + \text {OPT}(I).\) On the other hand, it is well known that \(\text {OPT}(I) \le \frac{n+1}{2} \cdot \sum _{j=1}^n p_j\), which is the expected objective value when scheduling non-preemptively in a random order, thus, \(\text {OPT}\le \frac{n+1}{2} (t+n)\). Therefore, we have \( \frac{\mathfrak {D}(I)}{\text {OPT}(I)} \ge 1+\frac{2nt}{(t+n)(n+1)} \ge 1+\frac{2nT}{(T+n)(n+1)} =3-\varepsilon \).
If \(j_0\) is probed for \(\tau =\infty \), i.e., it is processed non-preemptively until its completion, then for each job \(j\ne j_0\) we set \(p_j {:}{=}1+Y_j(T)\). Denote by \(\text {OPT}'\) the optimal \(\text {SPT}\) cost for jobs \([n] \setminus \{ j_{0} \}\), and set \(p_{j_0}:= 10\cdot \text {OPT}'\). As \(j_0\) is the first job to complete in I, we clearly have \(\mathfrak {D}(I) \ge n\cdot p_{j_0}= 10n \cdot \text {OPT}'\). On the other hand, \(\text {OPT}\) processes \(j_0\) last, so \(\text {OPT}(I)=\text {OPT}' + \sum _{j\ne j_0} p_j + p_{j_0} \le (1+1+10) \cdot \text {OPT}'\). This implies \(\frac{\mathfrak {D}(I)}{\text {OPT}(I)}\ge \frac{10n}{12}\ge 3-\frac{2}{n+1}\), where the last inequality holds for all \(n\ge 3\). \(\square \)
4 The b-scaling strategy
Let us now introduce the b-scaling algorithm \({\mathfrak {D}_{b}}\), which is the basis for most results in this paper. The idea of this algorithm is simple and quite natural: it proceeds by rounds \(q\in \mathbb {Z}\). In round \(q\) every non-completed job is probed (once) for \(w_j b^q\) in some prescribed order, where \(b>1\) is a constant. To execute \({\mathfrak {D}_{b}}\), we can store for each job its rank at time t, i.e., the largest \(q\) such that it was probed for \(w_jb^{q-1}\) until t. At any end of a probe, \({\mathfrak {D}_{b}}\) schedules the job j with minimum rank and minimum index for time \(w_jb^q\).
We also introduce a randomized variant of the algorithm. Randomization occurs in two places: First the jobs are reordered according to a random permutation \(\Sigma \) at the beginning of the algorithm. Second, we replace the probing time \(w_j b^q\) of the \(q\)th round with \(w_j b^{q+\Xi }\) for some random offset \(\Xi \in [0,1]\). Algorithm 1 gives the pseudocode of this strategy when it starts from round \(q_{0}\in \mathbb {Z}\), in which case it is denoted by \({\mathfrak {S}_{b}}^{\sigma ,\xi ,q_{0}}\). The kill-and-restart strategy \({\mathfrak {S}_{b}}^{\sigma ,\xi }\) studied in this paper can actually be seen as the limit of \({\mathfrak {S}_{b}}^{\sigma ,\xi ,q_{0}}\) when \(q_{0}\rightarrow -\infty \), and is described formally below. The deterministic b-scaling algorithm \({\mathfrak {D}_{b}}\) is obtained by setting \(\sigma ={\text {id}}\) (the identity permutation) and \(\xi =0\), while the randomized variant \({\mathfrak {R}_b}\) is obtained for a permutation \(\Sigma \) drawn uniformly at random from \(\mathcal {S}_n\) and a random uniform offset \(\Xi \sim \mathcal {U}([0,1])\), i.e.,
As for \(\text {OPT}\), by slight abuse of notation, we denote by \({\mathfrak {D}_{b}}(I)\) and \({\mathfrak {R}_b}(I)\) the schedule for instance I computed by \({\mathfrak {D}_{b}}\) and \({\mathfrak {R}_b}\), respectively, as well as its cost. We drop the dependence on I whenever the instance is clear from context.
While \({\mathfrak {S}_{b}}^{\text {id},0,q_{0}}\) can easily be implemented, it is not possible to implement the limit strategy \({\mathfrak {D}_{b}}\), for example, on a Turing machine, since at a time arbitrarily close to 0 it has probed each jobs an infinite number of times.
The b-scaling algorithm starting from a fixed round \(q_0\) is fully described by Algorithm 1. In practice, if a lower bound on the processing times is known (e.g. \(p_j \ge 1\)), it is enough to consider this implementable algorithm. We study the more general strategy \({\mathfrak {D}_{b}}\) for \(q_0 \rightarrow - \infty \) that we formally define in Appendix 8.
4.1 Tight analysis of the deterministic b-scaling strategy
In this section, we compute tight bounds for the competitive ratio of \({\mathfrak {D}_{b}}\) for . For the analysis, we need some additional definitions. We denote by \(s_j{:}{=}\frac{p_j}{w_j}\) the Smith ratio of job \(j\in [n]\). Further, we define \(D_{jk} {:}{=}Y_j^{{\mathfrak {D}_{b}}}(C_k^{{\mathfrak {D}_{b}}})\) as the amount of time spent probing job j before the completion of job k. For all \(j,k\in [n]\) we define the weighted mutual delay \(\Delta _{jk}\) by \(\Delta _{jk}{:}{=}w_j D_{jj}\) if \(j=k\) and \(\Delta _{jk}{:}{=}w_k D_{jk} + w_j D_{kj}\) if \(j\ne k\). Thus, it holds
Lemmas 3 to 5 constitute preparations for Theorem 6, establishing the upper bound on the competitive ratio. Afterwards, the tightness is proven in Theorem 7. The first step towards the upper bound is to provide an overestimator of \(\Delta _{jk}\) that is piecewise linear in \((s_j, s_k)\).
Lemma 3
Define the function \(F:\{ (s,s')\in \mathbb {R}_{>0}^2 \mid s\le s' \}\rightarrow \mathbb {R}\) by
Then F is non-decreasing in both arguments. Moreover, for all \(j,k\in [n]\) such that \(s_j\le s_k\), it holds that \(\Delta _{jk}\le w_j w_k\ F(s_j, s_k)\).
Proof
Let first \(s' > 0\) be fixed, and let \(r' {:}{=}\lfloor \log _b(s') \rfloor \). Then the function \(F(\cdot , s') :(0,s'] \rightarrow \mathbb {R}\) is obviously non-decreasing on \((0,b^{r'})\) and on \([b^{r'}, s']\). To see that it is also non-decreasing around the breakpoint \(b^{r'}\), we take the limit

Now let \(s > 0\) be fixed, and let \(r {:}{=}\lfloor \log _b(s) \rfloor \). Then the function \(F(s, \cdot ) :[s, \infty ) \rightarrow \mathbb {R}\) is clearly non-decreasing on \([s, b^{r+1})\) and on \([b^{r+1}, \infty )\). At the breakpoint we have

so that it is globally non-decreasing.
For all \(j\in [n]\), let \(q_j {:}{=}\lceil \log _b(s_j) \rceil \), so that \(b^{q_j-1} < s_j \le b^{q_j}\). We have \(D_{jj}=\sum _{i=-\infty }^{q_j-1} w_j b^i + p_j = w_j(\frac{b^{q_j}}{b-1}+s_j)\), so it holds \(\Delta _{jj} = w_j D_{jj} = w_j^2(\frac{b^{q_j}}{b-1}+s_j) \le w_j^2 F(s_j,s_j)\), where we have used the fact that \(q_j=\lceil \log _b(s_j) \rceil \le \lfloor \log _b(s_j) \rfloor +1\).
Now, let \(j\ne k\) such that \(s_j \le s_k\). We first assume that jobs j and k complete in the same round, i.e., \(\lceil \log _b(s_j)\rceil =\lceil \log _b(s_k)\rceil \). If job k is executed first in this round, then we have \(D_{jk}=\sum _{i=-\infty }^{q_j-1} w_j b^i = w_j\frac{b^{q_j}}{b-1}\) and \(D_{kj}=\sum _{i=-\infty }^{q_j-1} w_k b^i+p_k = w_k \frac{b^{q_j}}{b-1}+p_k\), which gives
Similarly, if job j is completed first, we have \(D_{jk}=w_j \frac{b^{q_j}}{b-1}+p_j\) and \(D_{kj}=w_k\frac{b^{q_j}}{b-1}\), so we obtain \(\Delta _{jk} = w_j w_k(\frac{2b^{q_j}}{b-1} + s_j) \le w_j w_k(\frac{2b^{\lfloor \log _b(s_j)\rfloor + 1}}{b-1} + s_k)\), i.e., the bound (1) is still valid. If \(\lfloor \log _b(s_j) \rfloor = \lfloor \log _b(s_k) \rfloor \), then the right-hand side equals \(w_j w_k F(s_j, s_k)\). Otherwise, \(s_k = b^{\lfloor \log _b(s_j) \rfloor + 1}\), so that

Now, assume that job k is completed in a later round than job j, i.e., \(\lceil \log _b(s_j)\rceil <\lceil \log _b(s_k)\rceil \). Then, \(D_{jk}=w_j \frac{b^{q_j}}{b-1}+p_j\) and \(D_{kj}\le w_k \frac{b^{q_j}}{b-1}+w_k b^{q_j}\), where the inequality is tight whenever job k is probed before job j in the round where j is completed. Thus,
If \(\lfloor \log _b(s_j) \rfloor < \lfloor \log _b(s_k) \rfloor \), then the right-hand side be bounded by \(w_j w_k F(s_j, s_k)\), using that \(q_j \le \lfloor \log _b(s_j) \rfloor + 1\). Otherwise, \(s_j = b^{q_j}< s_k < b^{q_j+1}\), so that

\(\square \)
Summing the bounds of the previous lemma yields
We next prove a lemma showing that for bounding the ratio \(U/\text {OPT}\) we can restrict to instances in which all Smith ratios are integer powers of b.
Lemma 4
For any instance \((\varvec{p},\varvec{w})\), there exists another instance \((\varvec{p}',\varvec{w})\) with \(p_j'=w_j b^{q_j}\) for some \(q_j\in \mathbb {N}_0\), for all \(j\in [n]\), such that
Proof
Let \(\rho _{\min } = \min _{j \in [n]} \lfloor \log _b(s_j) \rfloor \). Then in the instance \((b^{-\rho _{\min }} \varvec{p}, \varvec{w})\) all jobs have Smith ratio \(\ge 1\), and it holds that \(U(b^{-\rho _{\min }} \varvec{p}, \varvec{w}) = b^{-\rho _{\min }} \cdot U(\varvec{p}, \varvec{w})\). Clearly, we also have \(\text {OPT}(b^{-\rho _{\min }} \varvec{p}, \varvec{w}) = b^{-\rho _{\min }} \cdot \text {OPT}(\varvec{p}, \varvec{w})\). Therefore, without loss of generality, we assume that \(s_j \ge 1\) for all \(j \in [n]\). Let \(S {:}{=}\bigl \{ \log _b(s_j) \mid j\in [n]\bigr \}{\setminus } \mathbb {Z}\). If \(S=\emptyset \), we are done. Otherwise, let \(q {:}{=}\min (S) > 0\) and \(I{:}{=}\{ j\in [n] \mid s_j = b^q \}\). We will either decrease the Smith ratio of each job \(j\in I\) to \(b^{\lfloor q \rfloor }\) or increase them to \(b^{q'}\), where \(q' {:}{=}\min \bigl (S \cup \{\lceil q \rceil \} {\setminus } \{q\}\bigr )\), so that the cardinality of S is decreased by 1, and repeat this operation until each Smith ratio is an integer power of b. For \(\delta \in \mathbb {R}\), define \(p_j(\delta ) {:}{=}p_j+ \delta \cdot w_j \mathbb {1}_I(j)\), so the Smith ratio of job j in the instance \((\varvec{p}(\delta ),\varvec{w})\) is \(s_j(\delta )=s_j + \delta \) if \(j\in I\) and \(s_j(\delta )=s_j\) otherwise. Let \(\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt}{:}{=}b^{\lfloor q \rfloor }-b^q<0\) and \(\bar{\delta }{:}{=}b^{q'}-b^q>0\). Since an optimal schedule follows the \(\text {WSPT}\) rule, it is easy to see that the function \(\delta \mapsto \text {OPT}(\varvec{p}(\delta ),\varvec{w})\) is linear in the interval \([\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt},\bar{\delta })\), as the order of the Smith ratios remains unchanged for all \(\delta \) in this interval. For the same reason, and because for all j, k the Smith ratios \(s_j(\delta ),s_k(\delta )\) remain in the same piece of the piecewise linear function \((s_j,s_k) \mapsto F(\min (s_j,s_k), \max (s_j,s_k))\) for all \(\delta \in [\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt},\bar{\delta })\), the function \(\delta \mapsto U(\varvec{p}(\delta ),\varvec{w})\) is also linear over \([\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt},\bar{\delta })\). As a result, the function
is a quotient of linear functions and thus monotone over \([\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt},\bar{\delta })\). Indeed, \(\text {OPT}(\varvec{p}(\delta ),\varvec{w})>0\) for all \(\delta \ge \underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt}\), so h has no pole in this interval. We can thus distinguish two cases: if the function h is non-increasing, we let \(\delta '=\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt}\), so we have \(h(\delta ')\ge h(0)\), which means that we can decrease the Smith ratio of each job \(j\in I\) to \(s_j(\delta ')=s_j+\underline{\delta }\hspace{-1.66656pt}\hspace{1.66656pt}=b^q+b^{\lfloor q\rfloor }-b^q=b^{\lfloor q\rfloor } \ge 1\) without decreasing the bound \(U/\text {OPT}\) on the competitive ratio. Otherwise, the function h is non-decreasing, hence , where the last inequality comes from the fact that F is non-decreasing. So in this case we set \(\delta '=\bar{\delta }\) and we can increase the Smith ratio of each job \(j\in I\) to \(b^{q'}\) without decreasing the bound on the competitive ratio; if \(q'=\lceil q \rceil \), it means that we round up these Smith ratios to the next integer power of b, otherwise it is \(q'=\min (S\setminus \{q\})\), so we cluster together a larger group of jobs with a Smith ratio \(s_j(\bar{\delta })=b^{q'}\) that is not an integer power of b.
In all cases, the number of distinct non-integer values of \(\log _b(p_j(\delta ')/w_j)\) is decremented by one compared to the original instance, while the bound on the competitive ratio is only larger:
Repeating this construction until all Smith ratios are integer powers of b yields the desired result. \(\square \)
The next lemma gives a handy upper bound for the competitive ratio of \({\mathfrak {D}_{b}}\) relying on the ratio of two quadratic forms. For \(L\in \mathbb {N}_0\) define the symmetric \(((L+1) \times (L+1))\)-matrices \(\varvec{A}_L {:}{=}\big ( \frac{1}{2} b^{\min (\ell ,m)} \mathbb {1}_{\{\ell \ne m\}}\big )_{0\le \ell ,m \le L}\) and \(\varvec{B}_L {:}{=}\big ( \frac{1}{2} b^{\min (\ell ,m)}\big )_{0\le \ell ,m \le L}\).
Lemma 5
For any instance \((\varvec{p},\varvec{w})\) there exists an integer L and a vector \(\varvec{x}\in \mathbb {R}^{\{0,\dotsc ,L\}}\) such that
Proof
Consider an arbitrary instance \((\varvec{p},\varvec{w})\). It follows from (2) that \(\frac{{\mathfrak {D}_{b}}(\varvec{p},\varvec{w})}{\text {OPT}(\varvec{p},\varvec{w})}\le \frac{U(\varvec{p},\varvec{w})}{\text {OPT}(\varvec{p},\varvec{w})} \). By Lemma 4, we construct an instance \((\varvec{p}',\varvec{w})\) in which each Smith ratio is a non-negative integer power of b, and such that \(\frac{U(\varvec{p},\varvec{w})}{\text {OPT}(\varvec{p},\varvec{w})}\le \frac{U(\varvec{p}',\varvec{w})}{\text {OPT}(\varvec{p}',\varvec{w})} \) holds. In the remainder of this proof, we relabel the jobs so that \(\frac{p_1'}{w_1}\le \cdots \le \frac{p_n'}{w_n}\). We define \(L {:}{=}\max _{j \in [n]} \log _b(s'_j)\). For all \(\ell =0,\dotsc ,L\), we denote by \(J_\ell {:}{=}\{ j\in [n] \mid p_j'=w_jb^\ell \}\) the subset of jobs with Smith ratio equal to \(b^\ell \), so by construction we have \([n]=J_0\cup \cdots \cup J_L\). We also define \(x_\ell {:}{=}\sum _{j\in J_\ell } w_j\) and \(y_\ell {:}{=}\sum _{j\in J_\ell } w_j^2\), for all \(\ell =0,\dotsc ,L\).
We first get a handy expression for \(\text {OPT}(\varvec{p}',\varvec{w})\) relying on the vectors \(\varvec{x},\varvec{y}\in \mathbb {R}^{\{0,\dotsc ,L\}}\). By optimality of the \(\text {WSPT}\) rule,
On the other hand, using the fact that \(F(b^\ell ,b^m){=} b^\ell F(1, b^{m-\ell }){=}b^\ell F(1,b^{\min (m-\ell ,1)})\) for all integers \(\ell \le m\), we obtain
Substituting \(F(1,1)=\frac{2b}{b-1}+1\) and \(F(1,b)=\frac{2b}{b-1}+b+1\), we get
\(\square \)
In order to determine an upper bound for the competitive ratio of \({\mathfrak {D}_{b}}\), we need to bound the last term in the expression from Lemma 5. The latter is the ratio of two quadratic forms, and an upper bound for this term can be derived by computing the maximum eigenvalue of the matrix \(\varvec{Z}_L {:}{=}\varvec{Y}_L^{-\top } \varvec{A}_L \varvec{Y}_L^{-1}\), where \(\varvec{B}_L = \varvec{Y}_L^\top \varvec{Y}_L\) is the Cholesky decomposition of the matrix \(\varvec{B}_L\). An explicit computation of the matrix \(\varvec{Z}_L\) reveals that it is a tridiagonal matrix whose principal submatrix—obtained by deleting the first row and first column—is a (tridiagonal) Toeplitz matrix that we refer to as \(\varvec{T}_{L}\). Finding an upper bound for the largest eigenvalue of \(\varvec{Z}_L\) is the main ingredient of the proof of Theorem 6, while the eigenvector corresponding to this eigenvalue can be used to construct instances that prove the tightness of the bound (Theorem 7).
Theorem 6
For \(b>1\), \({\mathfrak {D}_{b}}\) is \(\bigl ( 1+\frac{2b^{\nicefrac {3}{2}}}{b-1} \bigr )\)-competitive for . This ratio is minimized for \(b=3\), yielding a performance guarantee of \(1+3\sqrt{3}\approx 6.196\).
Proof of Theorem 6
By Lemma 5 we have
As described above, for every \(L \in \mathbb {N}\),
is the maximum eigenvalue of the matrix \(\varvec{Z}_L = \varvec{Y}_L^{-\top } \varvec{A}_L \varvec{Y}_L^{-1}\). For \(\alpha , \beta \in \mathbb {R}\) let \(\varvec{T}_L(\alpha , \beta ) \in \mathbb {R}^{L \times L}\) denote the symmetric tridiagonal Toeplitz matrix with \(\alpha \) on the main diagonal and \(\beta \) on both adjacent diagonals. The explicit representation of \(\varvec{Z}_L\) can be derived by applying Lemma 25 in the appendix with \(a_1 = \cdots = a_L = 1\). As many terms in the general form cancel out, this yields the tridiagonal matrix
We now want to show that \(\lambda _{\max }(\varvec{Z}_L) \le \frac{2 (\sqrt{b} - 1)}{b-1}\). This is equivalent to the matrix \(\varvec{H} {:}{=}\frac{2 (\sqrt{b} - 1)}{b-1} \varvec{I}_{L+1} - \varvec{Z}_{L}\) being positive semidefinite, where \(\varvec{I}_{L+1}\) denotes the identity matrix with indices \(\{0,\dotsc ,L\}\). We compute
This matrix has the form required in Lemma 27 with \(k=1\), \(\alpha = 2-\frac{2}{\sqrt{b}}\), and \({\varvec{v} = -\sqrt{1-\frac{1}{b}} \in \mathbb {R}^1}\). Since \(\alpha - \Vert \varvec{v} \Vert ^2 = 2 - \frac{2}{\sqrt{b}} - \bigl (1 - \frac{1}{b}\bigr ) = \frac{(\sqrt{b} - 1)^2}{b} \ge 0\), the lemma implies that \(\varvec{H}\) is positive semidefinite, so that \(\lambda _{\max }(\varvec{Z}_L) \le \frac{2(\sqrt{b} -1)}{b-1}\). Since this holds for every \(L \in \mathbb {N}\), we obtain with inequality (5)
The latter is minimized for \(b=3\) yielding the performance guarantee of \(1 + 3 \sqrt{3} \). \(\square \)
Next, we show that our analysis of \({\mathfrak {D}_{b}}\) is asymptotically tight.
Theorem 7
For every \(b > 1\), there exists a sequence of instances \((\varvec{p}_L)_{L \in \mathbb {N}}\) for such that

Proof
For \(L \ge 1\) let \(\varvec{Y}_L\), and \( \varvec{Z}_{L}\) be the matrices defined above, and let \(\varvec{T}_L {:}{=}\varvec{T}_L(-\frac{2}{b-1}, \frac{\sqrt{b}}{b-1})\) be the principal submatrix of \(\varvec{Z}_L\) and \(\varvec{z}_L = (z^{(L)}_\ell )_{0 \le \ell \le L}\) with \(z^{(L)}_{\ell } {:}{=}\sqrt{\frac{2}{L+1}} \cdot \sin \bigl (\frac{\ell \pi }{L + 1}\bigr )\). By [51, Theorem 2.4], \(\tilde{\varvec{z}}_{L} {:}{=}(z^{(L)}_{\ell })_{1 \le \ell \le L} \) is the eigenvector of the matrix \(\varvec{T}_{L}\) corresponding to the largest eigenvalue \(\lambda _{\max } (\varvec{T}_{L}) = - \frac{2}{b-1} + 2\frac{\sqrt{b}}{b-1} \, \cos \bigl (\frac{\pi }{L+1} \bigr )\), and we have \(\Vert \varvec{z}_L \Vert ^2 = \Vert \tilde{\varvec{z}}_{L} \Vert ^2 = \frac{2}{L+1} \sum _{\ell =1}^{L} \sin ^2 \big (\frac{\ell \pi }{L+1}\big ) = 1\). Define \(\varvec{x}_L = (x_\ell ^{(L)})_{0 \le \ell \le L} {:}{=}\varvec{Y}_L^{-1} \varvec{z}_L\). By construction, it holds
The idea is to define for every \(L \in \mathbb {N}\) an instance \(\varvec{p}_L\) via a non-negative integer vector \(\varvec{n}_L \in \mathbb {N}_0^{L}\) that is similar to \(\varvec{x}_L\), which contains \(n_{\ell }\) jobs with processing time \(b^{\ell } + \varepsilon \) for all \(\ell \in [L]\) and for some \(\varepsilon >0\). There is an \(\ell ^* \in \mathbb {N}_{>0}\) such that \(x^{(L)}_{\ell } \ge 0\) holds for all \(L \ge \ell \ge \ell ^*\). This follows from Lemma 26 in the appendix because the matrix \(\varvec{Y}\) has exactly the required form, as shown in Lemma 24. Therefore, for \(L \ge \ell ^*\) the vector \(\varvec{n}_L = (n^{(L)}_{\ell })_{0 \le \ell \le L}\) with \(n_\ell ^{(L)} {:}{=}0\) for \(\ell < \ell ^*\) and \(n^{(L)}_\ell {:}{=}\lfloor b^L x^{(L)}_\ell \rfloor \) for \(\ell \ge \ell ^*\) is a non-negative integer vector, so that for every \(\varepsilon \ge 0\) the instance \(\varvec{p}_L(\varepsilon ) = (p_j^{(L)}(\varepsilon ))\) consisting of \(n_\ell ^{(L)}\) jobs with processing time \(b^\ell + \varepsilon \) for \(\ell = 0,\dotsc ,L\), ordered non-increasingly by processing times, is well-defined. Let \(n^{(L)} {:}{=}\sum _{\ell =0}^L n^{(L)}_\ell \) be the number of jobs in \(\varvec{p}_L(\varepsilon )\).
Let now \(\varepsilon > 0\). Clearly, we have
For every job j let \(q_j {:}{=}\lceil \log _b(p_j^{(L)}) \rceil \), i.e., \(q_j = \ell + 1\) for the \(n_\ell \) jobs with processing time \(b^\ell + \varepsilon \). Proceeding similarly as in the proof of Lemma 3, we obtain \(\Delta _{jj}=\frac{b^{q_j}}{b-1}+p_j^{(L)}=b^{q_j-1} (\frac{b}{b-1}+1) + \varepsilon \), and for \(j \ne k\) with \(q_j \le q_k\) we have \(\Delta _{jk}=\frac{2b^{q_j}}{b-1}+p_j^{(L)}=b^{q_j-1}(\frac{2b}{b-1}+1)+\varepsilon \) if \(q_j=q_k\) or \(\Delta _{jk}=\frac{2b^{q_j}}{b-1}+b^{q_j}+p_j^{(L)}=b^{q_j-1}(\frac{2b}{b-1}+b+1)+\varepsilon \) otherwise. This can be rewritten as
Summing over all pairs of jobs, we thus obtain
As the processing times of \(\varvec{p}_L(0)\) are integer powers of b, we can use Equations (4) and (3) with \(\varvec{y}=\varvec{x}=\varvec{n}_L\), resulting in
In the following we compute the limit for \(L \rightarrow \infty \) by computing the limits of the occurring terms separately. By Lemma 26, \(\sum _{\ell =\ell ^*}^L x_\ell ^{(L)} \xrightarrow {L\rightarrow \infty } 0\). Therefore, we have
For \(L \ge \ell ^*\) let \(\varvec{x}_1^{(L)} = (x_\ell ^{(L)})_{0 \le \ell \le \ell ^*-1}\), \(\varvec{x}^{(L)}_2 = (x_\ell ^{(L)})_{\ell ^* \le \ell \le L}\), \(\varvec{A}_{11}^{(L)} = (A_{\ell m}^{(L)})_{0 \le \ell , m \le \ell ^* - 1}\), \(\varvec{A}_{12}^{(L)} = (A_{\ell m}^{(L)})_{\begin{array}{c} 0 \le \ell \le \ell ^*-1\\ \ell ^* \le m \le L \end{array}}\), and \(\varvec{A}_{22}^{(L)} = (A_{\ell m}^{(L)})_{\ell ^* \le \ell ,m \le L}\), so that
and let \(\varvec{n}_2^{(L)} = (n_\ell ^{(L)})_{\ell ^* \le \ell \le L}\). With the definition of \(\varvec{n}_L\), we compute
For the first summand we have
and for the second summand, by Lemma 26, we have
Finally, the third summand satisfies
Similarly, we have
As \(\varvec{x}_L^\top \varvec{B}_L \varvec{x}_L = \Vert \varvec{z}_L \Vert = 1 \ne 0\) for all L, we have thus shown that
By Lemma 28 there is a sequence of problem instances for which the competitive ratio converges to the right-hand side. \(\square \)
4.2 Tight analysis of the randomized b-scaling strategy
We now consider \({\mathfrak {R}_b}= {\mathfrak {S}_{b}}^{\Sigma , \Xi }\), where \(\Sigma \) is a permutation drawn uniformly at random from \(\mathcal {S}_n\) and \(\Xi \) is uniformly distributed on the interval [0, 1].
As in the analysis of the deterministic algorithm, we start with a lemma giving an overestimator of \(\mathbb {E}[\Delta _{jk}]\) for jobs j and k such that \(s_j\le s_k\). This time, our overestimator is not piecewise linear in \(s_j\) and \(s_k\) anymore. Instead, it depends on a concave function applied to the ratio \(\frac{s_k}{s_j}\ge 1\).
Lemma 8
Let \(f :[1,b] \rightarrow \mathbb {R}\) be defined by
Then f is positive and increasing, and for all \(j\ne k\) such that \(s_j\le s_k\) it holds
Proof
By straightforward calculus, we obtain
Since \(f''(\alpha ) < 0\), the function \(\alpha \mapsto f'(\alpha )\) is decreasing over [1, b]. Hence, for all \(\alpha \in [1,b]\) we have \(f'(\alpha )\ge f'(b)=\frac{1}{2b\ln b} > 0\), which proves that f is increasing. Therefore, \(f(\alpha ) \ge f(1) = 1 + \frac{2}{\ln b} > 0\), concluding the proof of the first part of the lemma.
Let now \(j, k \in [n]\), \(j \ne k\) with \(s_j \le s_k\) be fixed, and let \(\ell \in \mathbb {Z}\) and \(u \in (0,1]\) be such that \(s_j=b^{\ell +u}\). Moreover, let \(q_j(\Xi ) {:}{=}\lceil \log _b(s_j) - \Xi \rceil = \ell + \mathbb {1}_{\{\Xi < u\}}\) be such that j completes in the round, where jobs are probed for \(w_j b^{q_j(\Xi ) + \Xi }\). We thus have
Therefore, we get
Now, we fix a realization \(\xi \) of \(\Xi \) and compute \(\mathbb {E}[\Delta _{jk}\vert \Xi =\xi ]\) for another job k such that \(s_j\le s_k\). If \(q_j(\xi )=q_k(\xi )\), then \(\mathbb {E}[D_{jk}\vert \Xi =\xi ]=w_j \frac{b^{q_j(\xi )+\xi }}{b-1}+\frac{1}{2}\cdot p_j\) and \(\mathbb {E}[D_{kj}\vert \Xi =\xi ]=w_k \frac{b^{q_j(\xi )+\xi }}{b-1}+\frac{1}{2}\cdot p_k\), where the factors \(\frac{1}{2}\) in front of \(p_j\) and \(p_k\) come from the fact that job j is completed with probability \(\frac{1}{2}\) before job k due to the random permutation of the jobs. Otherwise, it is \(q_j(\xi )<q_k(\xi )\) and we have \(\mathbb {E}[D_{jk}\vert \Xi =\xi ]=w_j \frac{b^{q_j(\xi )+\xi }}{b-1}+p_j\), \(\mathbb {E}[D_{kj}\vert \Xi =\xi ]=w_k \frac{b^{q_j(\xi )+\xi }}{b-1}+w_k \frac{b^{q_j(\xi )+\xi }}{2}\). Putting all together,
Let us first consider the case \(s_k\ge b \cdot s_j\), as in this case \(q_j(\xi )<q_k(\xi )\) for all \(\xi \in (0,1)\). Thus,
It remains to handle the case \(s_j \le s_k < b \cdot s_j\), in which it can occur that jobs j and k are completed in the same round. Let \(\delta \in [0,1)\) such that \(s_k=s_j\cdot b^{\delta }=b^{\ell +u+\delta }\). To compute \(\mathbb {E}[\Delta _{jk}]\), we have to distinguish between the cases \(u+\delta \le 1\) and \(u+\delta > 1\). We only handle the former case, as the latter can be handled similarly and yields the same formula, so in the remainder of this proof we assume \(u+\delta \le 1\). Then, it holds \(q_k(\xi )=\ell +\mathbb {1}_{\{\xi <u+\delta \}}\), so that \((q_j(\xi ),q_k(\xi ))=(\ell +1,\ell +1)\) if \(\xi \in [0,u)\), \((q_j(\xi ),q_k(\xi ))=(\ell ,\ell +1)\) if \(\xi \in [u,u+\delta )\), and \((q_j(\xi ),q_k(\xi ))=(\ell ,\ell )\) if \(\xi \in [u+\delta ,1]\). Thus, we can write
\(\square \)
The expressions of \(\mathbb {E}[\Delta _{jk}]\) derived in the previous lemma show that \({\mathfrak {R}_b}(\alpha \varvec{p},\varvec{w})=\alpha {\mathfrak {R}_b}(\varvec{p},\varvec{w})\) holds for all \(\alpha >0\). Since the same trivially holds for the optimal solution, i.e., \(\text {OPT}(\alpha \varvec{p},\varvec{w})=\alpha \text {OPT}(\varvec{p},\varvec{w})\) we can assume without loss of generality that the instance has been rescaled, so that \(\min _{j\in [n]} s_j=1\). Moreover, we relabel the jobs so that \(s_1 \le \cdots \le s_n\). Then, summing the bounds from the previous lemma yields
In order to obtain an upper bound, we use a similar technique as in Sect. 4.1. However, the proof is more involved because the bound from Lemma 8 is not piecewise linear in \(s_j\) and \(s_k\), so we cannot construct a worse instance \((\varvec{p}',\varvec{w})\) in which all Smith ratios are integer powers of b. Instead, we are going to subdivide the Smith ratios in intervals of the form \([b^{i/K},b^{(i+1)/K})\) for some integer K, and we will get a bound by grouping all jobs in an interval. As in Theorem 6, this bound involves a ratio of two quadratic forms (Lemma 9), but this time the maximization of this fraction amounts to finding the maximum eigenvalue of a banded Toeplitz matrix of bandwidth \(2K-1\).
Let \(K \in \mathbb {N}_{>0}\) and \(\beta {:}{=}b^{\frac{1}{K}}\). For \(L \in \mathbb {N}\) define the symmetric matrices \(\varvec{A}_L {:}{=}(\frac{1}{2} a_{\vert m-\ell \vert } \beta ^{\min (\ell ,m)})_{0 \le \ell , m \le L}\), where \(a_i {:}{=}f(\beta ^{\min (K,i+1)}) - f(\beta )\) for \(i \in \{0,\dotsc ,L\}\), and let \(\varvec{B}_L {:}{=}(\frac{1}{2} \beta ^{\min (\ell ,m)})_{0 \le \ell , m \le L}\).
Lemma 9
For any instance \((\varvec{p}, \varvec{w})\) there exists \(L \in \mathbb {N}\) and a vector \(\varvec{x} \in \mathbb {R}^{\{0,\dotsc ,L\}}\) such that
Proof
Let L be a multiple of K! such that \(s_j \le \beta ^L = b^{L/K}\) for all \(j \in [n]\). For all \(\ell \in \{0,\dotsc ,L\}\) we define \(J_\ell {:}{=}\{j \in [n] \mid \beta ^{\ell }\le s_j < \beta ^{\ell +1}\}\), \(x_\ell {:}{=}\sum _{j\in J_\ell } w_j\), and \(y_\ell {:}{=}\sum _{j\in J_\ell } w_j^2\). Then \([n] = J_0 \cup \cdots \cup J_L\). We obtain as a lower bound on the optimal cost
On the other hand, using (8), we compute

Consequently,
\(\square \)
We next prove the main result of this section.
Theorem 10
For every \(b>1\), \({\mathfrak {R}_b}\) is \(\frac{2b+\sqrt{b}-1}{\sqrt{b} \ln b}\)-competitive for . This ratio is minimized for \(b\approx 8.16\), yielding a performance guarantee smaller than 3.032.
Proof
By Lemma 9 we have for every \(K \in \mathbb {N}_{>1}\) and \(\beta = b^{\frac{1}{K}}\) that
For now let \(K \in \mathbb {N}_{>1}\) be fixed. Similarly to the proof of Theorem 6, Lemma 25 yields that the inner supremum is
for the matrix

where \(\varvec{u}\in \mathbb {R}^{K-1}\) has coordinates \(u_i=\frac{f(\beta ^{i+1})-f(\beta ^i)}{\sqrt{\beta ^i-\beta ^{i-1}}}\), \(i=1,\ldots ,K-1\), and \(\varvec{T}_L\) is the \(L\times L\) banded symmetric Toeplitz matrix of bandwidth \(2(K-1)+1\) with elements
on its kth and \(-k\)th superdiagonals (where for \(k\in \mathbb {Z}\), the kth superdiagonal is the set of coordinates \((\varvec{T}_L)_{\ell ,m}\) such that \(m-\ell =k\); in particular, the 0th superdiagonal corresponds to the main diagonal). Substitution of \(f(\beta ^k)\) with its value \(\frac{1}{2} \bigl (1+\frac{k}{K} + \beta ^k \bigl (1 - \frac{k}{K}\bigr ) + \frac{\beta ^k+3}{\ln b}\bigr )\) yields the following simplified expression for \(t_k\):
We next show that \(\lambda _{\max }(\varvec{Z}_L)\le t_0 + 2\sum _{k=1}^{K-1} t_k\). To this end, we form the matrix
and prove that this matrix is positive semidefinite. For every \(i \in [K-1]\) the sum \(\sum _{k=i}^{K-1} \frac{t_k}{\beta ^{k/2}}\) is a telescoping sum, which sums up to \(\frac{f(\beta ^{i+1}) - f(\beta ^i)}{\beta ^i - \beta ^{i-1}} = \frac{u_i}{\sqrt{\beta ^i-\beta ^{i-1}}}\). In particular, \(\sum _{k=1}^{K-1} \frac{t_k}{\beta ^{k/2}} = \frac{f(\beta ^2)-f(\beta )}{\beta -1}\), so that \(t_0 = -2 \sum _{k=1}^{K-1} \frac{t_k}{\beta ^{k/2}}\). Therefore, we can rewrite \(\varvec{H}\) as a linear combination of \(t_1,\ldots ,t_{K-1}\):

\(\varvec{v}_k\in \mathbb {R}^k\) is a vector with coordinates \((\varvec{v}_k)_i= -\frac{\sqrt{\beta -1}}{\beta ^{(k-i+1)/2}}\), (\(i=1,\ldots ,k\)), and \(\varvec{T}_{k,L}\) is the sparse symmetric Toeplitz matrix of size \(L\times L\) whose only non-zero elements are 2 on the main diagonal and \(-1\) on the kth and \(-k\)th superdiagonals, i.e., \({(\varvec{T}_{k,L})_{ij}= 2\cdot \mathbb {1}_{\{i=j\}} - \mathbb {1}_{\{\vert i-j\vert =k\}}}\). To show that \(\varvec{H}\) is positive semidefinite, it suffices to show that \(t_k\ge 0\) and \(\varvec{H}_k\) is positive semidefinite for all \(k\in [K-1]\).
Let \(k\in [K-1]\). If \(k=K-1\), then \(t_{k}\ge 0\) follows from (10) and the fact that f is non-decreasing over [1, b]. Otherwise, this inequality follows from Equation (11). Next, we use Lemma 27 to show that \(\varvec{H}_k\) is positive semidefinite. This is possible because \(k\mid L\), as L is a multiple of K!, and
This concludes the proof that all \(\varvec{H}_k\) are positive semidefinite, hence \(\lambda _{\max }(\varvec{Z})\le t_0+2\sum _{k=1}^{K-1} t_k\). Together, we have shown that
for every \(K \in \mathbb {N}\).
In the final part of the proof we show that the right-hand side converges to \(\frac{2b + \sqrt{b}- 1}{\sqrt{b} \ln b}\) for \(K \rightarrow \infty \). To this end, let us now compute the sum
Moreover, we have \(t_0\xrightarrow {K\rightarrow \infty }-(1+\frac{1}{\ln b})\) and \(t_{K-1}=\frac{1}{2} b^{-\frac{K-1}{2K}}\bigl (\frac{b}{\ln b}-\frac{b-b^{\frac{1}{K}}}{K(b^{\frac{1}{K}}-1)}\bigr ) \xrightarrow {K\rightarrow \infty } \frac{1}{2\sqrt{b}}(\frac{b}{\ln b}-\frac{b-1}{\ln b})=\frac{1}{2\sqrt{b}\ln b}\). By using \(\beta \xrightarrow {K\rightarrow \infty } 1\) and \(f(\beta )\xrightarrow {K\rightarrow \infty } f(1)=1+\frac{2}{\ln b}\), we obtain the final bound by taking the limit when \(K\rightarrow \infty \):
A numerical minimization yields an optimal value of \(b \approx 8.16\) with a performance guarantee smaller than 3.032. \(\square \)
We next show that our analysis is tight.
Theorem 11
For all \(b>1\), there exists a sequence of instances \((\varvec{p}_L)_{L\in \mathbb {N}}\) for such that
Proof
For \(K \in \mathbb {N}\) let \(\beta = b^{1/K}\), and for \(L \in \mathbb {N}\) let \(\varvec{A}_L' {:}{=}(\frac{1}{2} a_{\vert m-\ell \vert }' \beta ^{\min (\ell ,m)})_{0 \le \ell , m \le L}\), where \(a_i' {:}{=}f(\beta ^{\min (K,i)}) - f(1)\) for \(i \in \{0,\dotsc ,L\}\). Moreover, let \(\varvec{B}_L = \varvec{Y}_L^\top \varvec{Y}_L\) be the Cholesky decomposition of \(\varvec{B}_L\). By Lemma 25, we have

for some \(\varvec{u}'\in \mathbb {R}^{K}\), where \(\varvec{T}'_L\) is the \(L\times L\) banded symmetric Toeplitz matrix of bandwidth \(2K+1\) with elements
on the kth and \(-k\)th superdiagonals. As in the proof of Theorem 7 for the deterministic version of the strategy, define \(\varvec{z}_L = (z_\ell ^L)_{0 \le \ell \le L}\) with \(z_{\ell }^{(L)} {:}{=}\sqrt{\frac{2}{L+1}} \cdot \sin \big ( \frac{\ell \pi }{L+1} \big )\) for \(\ell =0,\dotsc ,L\), and \(\varvec{x}_L {:}{=}\varvec{Y}_L^{-1} \varvec{z}_L \in \mathbb {R}^{\{0,\dotsc ,L\}}\). By construction,
where \(\tilde{\varvec{z}}{:}{=}[z_1,z_2,\ldots ,z_{L}]^\top \in \mathbb {R}^L\) and we have used the fact that \(z_0=0\) for the last equality. Furthermore,
Unlike the proof of Theorem 7 however, it is not true anymore that \(\tilde{\varvec{z}}\) is the eigenvector corresponding to the largest eigenvalue of \(\varvec{T}'_L\) because \(\varvec{T}'_L\) is not tridiagonal.
Lemmas 24 and 26 imply that there is an \(\ell ^* \in \mathbb {N}_{>0}\) such that \(x_\ell ^{(L)} \ge 0\) for all \(L \ge \ell \ge \ell ^*\). Therefore, for \(L \ge \ell ^*\) the vector \(\varvec{n}_L = (n^{(L)}_{\ell })_{\ell =0}^L\) with \(n_\ell ^{(L)} {:}{=}0\) for \(\ell < \ell ^*\) and \(n^{(L)}_\ell {:}{=}\lfloor \beta ^L x^{(L)}_\ell \rfloor \) for \(\ell \ge \ell ^*\) is a non-negative integer vector, defining the instance \(\varvec{p}_L\) that consists of \(n_\ell ^{(L)}\) jobs with processing time \(\beta ^\ell \) for \(\ell = 0,\dotsc ,L\).
For such an instance, we have
and
Then, proceeding similarly as in the proof of Lemma 5, we obtain
Therefore,
As in the proof of Theorem 7, we compute the limits of the occurring terms for \(L \rightarrow \infty \).
where the convergence follows from Lemma 26. Thus, we see that the last summands of numerator and denominator go towards zero. Next, using that for \(\ell \ne m\) the absolute value of the entry of \(\varvec{A}_L'\) indexed by \(\ell \) and m is bounded by \(\frac{1}{2} f(b) \beta ^{\min (\ell , m)}\), exactly the same calculation as in the proof of Theorem 7 shows that
and, analogously,
We are going to show that \(\tilde{\varvec{z}}_L^\top \varvec{T}'_L \tilde{\varvec{z}}_L\) converges to \(t_0'+2\sum _{k=1}^K t_k'\) as the dimension L grows to \(\infty \). For this, define the function \(\Phi :\theta \mapsto t_0' + 2\sum _{k=1}^K t_k'\cos (kx)\), which is the Fourier Series associated with the Toeplitz matrix \(\varvec{T}'_L\). We claim that for all \(\ell \in [L]\), it holds
To see this, we first extend the definition of \(\tilde{z}_{\ell }^{(L)} = \sqrt{\frac{2}{L+1}} \sin \big ( \frac{\ell \pi }{L+1} \big )\) to all \(\ell \in \{1-K,\ldots ,L+K\}\), and we observe that for all \(\ell \in [L]\) and \(k\in [K]\) we have
Then, we use the fact that \(\tilde{z}_\ell \le 0\) for all \(\ell \in \{1-K,\ldots ,0\} \cup \{L+1,\ldots ,L+K\}\), so we have
Consequently, using the fact that \(\tilde{z}_\ell ^{(L)}\) is positive for all \(\ell \in [L]\) we obtain
We have thus constructed for every \(K \in \mathbb {N}\) a sequence \(\varvec{p}_L\) for which the competitive ratio converges to \(t_0' + \sum _{k=1}^K t_k'\). Further, similar calculations as in the proof of Theorem 10 show that \(t_0'+2\sum _{k=1}^K t_k'\xrightarrow {K\rightarrow \infty } -1-\frac{1}{\ln b}+\frac{2b-1}{\sqrt{b}\ln b}\). By Lemma 28 there is a sequence of problem instances for which the competitive ratio of \({\mathfrak {R}_b}\) converges to the desired value of
\(\square \)
5 Weighted shortest elapsed time first
In this section we consider the online time model, where each job j arrives at its release date \(r_j\) and is not known before that time. Thus, an instance for our problem is now given by a triple \(I = (\varvec{p}, \varvec{w}, \varvec{r})\) of processing times, weights, and release dates of all jobs. We consider the classical Weighted Shortest Elapsed Time First (\(\text {WSETF}\)) rule for this model. Intuitively, \(\text {WSETF}\) is the limit for \(\varepsilon \rightarrow 0\) of the algorithm that divides the time into time slices of length \(\varepsilon \) and in each time slice processes a job with minimum ratio of elapsed processing time over weight. To formalize this limit process we allow fractional schedules \(\text {S}\) that, at every point in time t, assign each job j a rate \(y_j^{\text {S}}(t) \in [0,1]\) such that \(\sum _{j=1}^n y_j^{\text {S}}(t) \le 1\) for all \(t \in \mathbb {R}_{\ge 0}\) and \(y_j^{\text {S}}(t) = 0\) if \(t < r_j\) or \(t > C_j^{\text {S}}(I)\), where \(C_j^{\text {S}}(I)\) is the smallest t such that \(Y_j^{\text {S}}(I, t) {:}{=}\int _0^t y_j^{\text {S}}(s)\,\text {d} s \ge p_j\) (this requires \(y_j^{\text {S}}\) to be measurable). At any time t let J(t) be the set of all released and unfinished jobs, and let A(t) be the set of all jobs from J(t) that currently have minimum ratio of elapsed time over weight. Then \(\text {WSETF}\) sets the rate for all jobs \(j \in J(t)\) to

In other words, \(\text {WSETF}\) always distributes the available processor rate among the jobs in J(t) so as to maximize \(\min _{j \in J(t)} Y_j^{\text {WSETF}}(I, t)/w_j\). An example is given in Fig. 1.
The following theorem gives the tight competitive ratio of \(\text {WSETF}\) for non-clairvoyant online scheduling on a single machine.
Theorem 12
\(\text {WSETF}\) is 2-competitive for .
We start by collecting some simple properties of the schedule created by \(\text {WSETF}\).
Lemma 13
Consider an instance \(I = (\varvec{p}, \varvec{w}, \varvec{r})\), and let j, k be two jobs with \(r_k < C_j^{\text {WSETF}}(I)\) and \(p_k/w_k \le p_j/w_j\). Then \(C_k^{\text {WSETF}}(I) \le C_j^{\text {WSETF}}(I)\).
Proof
Suppose that \(C_j^{\text {WSETF}}(I) < C_k^{\text {WSETF}}(I)\). The job j must be processed at a positive rate during some interval \((t, C_j^{\text {WSETF}}(I))\) with \(t < C_j^{\text {WSETF}}(I)\) by the \(\text {WSETF}\) schedule, meaning that \(j \in A(C_j^{\text {WSETF}}(I))\). Hence, we get the contradiction
where the last inequality holds because k completes after \(C_j^{\text {WSETF}}(I)\). \(\square \)
The lemma implies that in an instance with trivial release dates, for which \(\text {WSETF}\) coincides with the Weighted Round-Robin algorithm, analyzed by Kim and Chwa [6], the jobs j are completed in the order of their Smith ratios \(p_j/w_j\). In this case the weighted delay of each job in the \(\text {WSETF}\) schedule compared to the optimal \(\text {WSPT}\) schedule is exactly its processing time multiplied with the total weight of jobs with larger index.
Lemma 14
Let \(I_0 = (\varvec{p},\varvec{w},\varvec{0})\) be an instance with trivial release dates and \(p_1/w_1 \le \cdots \le p_n/w_n\). For every job \(j \in [n]\) we have

Proof
This will be shown by induction on n. Clearly, the statement is true if there is only a single job. So in the following let \(n > 1\). We have
so the statement holds for the first job. In order to show the statement for all other jobs, we consider the problem instance \(I_0'\) with job set \(J' {:}{=}\{2,\dotsc ,n\}\). For every \(j \in J'\) it holds that \(C_j^{\text {WSPT}}(I_0) = p_1 + C_j^{\text {WSPT}}(I'_0)\). In the \(\text {WSETF}\) schedule for \(I_0\) every \(j \in J'\) is processed at a rate of \(w_j/\sum _{k=1}^n w_k\) until time \(C_1^{\text {WSETF}}(I_0)\), while in the \(\text {WSETF}\) schedule for \(I_0'\) it is first processed at a rate of \(w_j/\sum _{k=2}^n w_k\). Since
every job j has received the same amount of processing in the \(\text {WSETF}\) schedule for \(I_0\) at time \(C_1^{\text {WSETF}}(I_0)\) as in the \(\text {WSETF}\) schedule for \(I_0'\) at time \(C_1^{\text {WSETF}}(I_0) - p_1\). Thus, for any time \(t > C_1^{\text {WSETF}}(I_0)\) the \(\text {WSETF}\) schedule for \(I_0\) at time t coincides with the \(\text {WSETF}\) schedule for \(I_0'\) at time \(t-p_1\). Therefore, \(C_j^{\text {WSETF}}(I_0) = p_1 + C_j^{\text {WSETF}}(I_0')\) for all jobs \(j \in J'\). Putting things together, we obtain for all \(j\in J'\)
where we applied the induction hypothesis to the instance \(I_0'\) with \(n-1\) jobs. \(\square \)
To bound the optimum objective value from below, we consider the mean busy times
of the jobs j in an arbitrary schedule \(\text {S}\). Since the mean busy time of each job is smaller than its completion time, the sum of weighted mean busy times is a lower bound on the sum of weighted completion times. It is well known [52, 53] that the former is minimized by the Preemptive WSPT (\(\text {PWSPT}\)) rule, which always processes the available job with smallest index (i.e. an available job with smallest Smith ratio \(p_j/w_j\)). Thus, the sum of weighted mean busy times in the \(\text {PWSPT}\) schedule is a lower bound on \(\text {OPT}(I)\), and it suffices to show that
The \(\text {PWSPT}\) rule is illustrated in Fig. 1. Note that the inequality holds with equality for instances \(I_0\) with trivial release dates because for such instances we have \(C_j^{\text {WSPT}} = M_j^{\text {PWSPT}} + p_j/2\), so that, by Lemma 14,
for every \(j \in [n]\). By summing over all jobs j we deduce
This implies the 2-competitiveness of the Weighted Round-Robin algorithm, proved by Kim and Chwa. In the remainder of this section this argument will be generalized to jobs released over time. We start by reducing the instance to a simpler case without changing the values to be compared.
Lemma 15
For every instance \(I = (\varvec{p}, \varvec{w}, \varvec{r})\) there is an instance \(I' = (\varvec{p}', \varvec{w}', \varvec{r}')\) consisting of \(n'\) jobs such that no job is preempted in the \(\text {PWSPT}\) schedule for \(I'\), \(\sum _{j=1}^{n'} w_j' C_j^{\text {WSETF}}(I') = \sum _{j=1}^n w_j C_j^{\text {WSETF}}(I)\), and \(\sum _{j=1}^{n'} w_j' M_j^{\text {PWSPT}}(I') = \sum _{j=1}^n w_j M_j^{\text {PWSPT}}(I)\).
Proof
We split every job into subjobs corresponding to the parts processed without interruption in the \(\text {PWSPT}\) schedule, i.e., we replace each job j by jobs \((j,1),\dotsc ,(j,\ell _j)\) such that \(\sum _{i=1}^{\ell _j} p_{(j,i)}' = p_j\). Moreover, we set the weights to \(w_{(j,i)} {:}{=}\frac{p_{(j,i)}}{p_j} \cdot w_j\), so that all parts have the same Smith ratio as the original job. Finally, the release dates are set to \(r_{(j,i)} {:}{=}r_j\). This operation does not change the sum of weighted mean busy times in the \(\text {PWSPT}\) schedule. Moreover, in the \(\text {WSETF}\) schedule, since all these jobs are released simultaneously and have the same Smith ratio, they will always be processed in a way so that their weighted elapsed times increase equally, so that they are all completed at the same time. Moreover, the total rate assigned to the jobs (j, i) for a fixed j equals the rate assigned to j in the original schedule. Therefore, all these jobs finish exactly at the time when job j is completed in the original schedule. \(\square \)
From now on we always consider an instance \(I = (\varvec{p}, \varvec{w}, \varvec{r})\) such that no job is preempted in the \(\text {PWSPT}\) schedule and \(p_1/w_1 \le \dotsc \le p_n/w_n\). Since the \(\text {PWSPT}\) and the \(\text {WSETF}\) strategies both fully utilize the machine whenever some jobs are available, the resulting schedules have the same idle intervals. By splitting into multiple independent subinstances, we can additionally assume that I has no idle intervals. We omit the instance in the notation for completion and elapsed times. For every fixed job j let \(\rho _j\) be the first point in time such that during \((\rho _j,C_j^{\text {WSETF}}]\) the machine continuously processes jobs k with \(Y_k^{\text {WSETF}}(C_j^{\text {WSETF}})/w_k \le p_j/w_j\), and let R(j) be the set of jobs processed in this interval. The times \(\rho _j\) are shown in the example in Fig. 1, and the sets R(j) consist exactly of the jobs released within \((\rho _j, C_j^{\text {WSETF}})\) because if a job \(k \in R(j)\) were released before \(\rho _j\), the machine would only process jobs l with \(Y_l^{\text {WSETF}}(\rho _j)/w_l \le Y_k^{\text {WSETF}}(\rho _j)/w_k \le p_j/w_j\) between \(r_k\) and \(\rho _j\), in contradiction to the minimality of \(\rho _j\). Let I(j) be the instance with job set R(j) where all jobs are released at time 0. In the next lemma we compare the completion time of j in the \(\text {WSETF}\) schedule for the original instance I to the corresponding completion time in the instance I(j).
Lemma 16
\(C_j^{\text {WSETF}} = \rho _j + C_j^{\text {WSETF}}(I(j))\) for every job j.
Proof
In the \(\text {WSETF}\) schedule for I every job \(k \in R(j)\) that is completed between the times \(\rho _j\) and \(C_j^{\text {WSETF}}(I)\) has \(p_k/w_k = Y_k^{\text {WSETF}}(C_j^{\text {WSETF}})/w_k \le p_j/w_j\). Hence, by Lemma 13, it also finishes before j in the \(\text {WSETF}\) schedule for I(j). All other jobs \(k \in R(j)\) have received processing time \(\frac{w_k}{w_j} \cdot p_j\) before the completion of j in both schedules. Therefore, the total processing that jobs from R(j) receive in the \(\text {WSETF}\) schedule for I between the times \(\rho _j\) and \(C_j^{\text {WSETF}}\) equals their total processing before the completion of j in the \(\text {WSETF}\) schedule for I(j). Since in both schedules the machine is continuously processing jobs from R(j) in the considered time, this implies the lemma. \(\square \)
Analogously to Lemma 16, we compare in the next lemma the preemptive WSPT schedules for I and I(j).
Lemma 17
For every job j
Proof
By Lemma 13, every job \(k \le j\) released before time \(C_j^{\text {WSETF}}\) has \(C_k^{\text {WSETF}} \le C_j^{\text {WSETF}}\). \(\text {PWSPT}\) always schedules some job \(k \le j\) at a rate of 1 if one is available. Thus, until every point in time, it cannot have spent less time on these jobs than \(\text {WSETF}\). Hence, at time \(C_j^{\text {WSETF}}\), \(\text {PWSPT}\) must have finished all these jobs as well. In particular, it has completed j. Therefore, we have \(C_j^{\text {PWSPT}} \le C_j^{\text {WSETF}}\) for every job j.
Every job \(k \in R(j)\) completed before \(C_j^{\text {PWSPT}}\) in the \(\text {PWSPT}\) schedule must be completely processed during \((\rho _j, C_j^{\text {PWSPT}}]\). Therefore,
\(\square \)
Now we are ready to prove the theorem.
Proof of Theorem 12
Using the fact that \(\text {PWSPT}\) is non-preemptive, we have \(C_j^{\text {PWSPT}} = M_j^{\text {PWSPT}} + \frac{p_j}{2}\) for all j, and thus
We will now bound the summands from the last sum by an expression generalizing (⁎) from Lemma 14. For a fixed \(j \in [n]\) Lemmas 16 and 17 yield
Since I(j) is an instance with trivial release dates, by Lemma 14,
where the first inequality holds because \(w_j p_k \le w_k p_j\) for \(k < j\) and \(w_j p_k \ge w_k p_j\) for \(k > j\). Now we sum over all jobs j, obtaining
Substituting this inequality into (18), we can generalize the computation in Equation (17):
\(\square \)
6 Extensions of the b-scaling strategy for release dates and parallel machines
In this section, we present extensions of the b-scaling strategies for the settings and
. We denote by \((\varvec{p},\varvec{w},\varvec{r},m)\) an instance on m identical parallel machines in which each job j has processing time \(p_j\), weight \(w_j\), and release date \(r_j\). In order to obtain bounds on the competitive ratios of these extensions, we compare the schedules of \({\mathfrak {D}_{b}}\) to schedules of algorithms that are constant-competitive. In particular, we use the 2-competitiveness of \(\text {WSETF}\) from the previous section for
and the 2-competitiveness of the round-robin (\(\text {RR}\)) schedule for
. We begin with the relation of the optimal costs of two instances whose processing times and release dates differ only by a multiplicative factor.
Lemma 18
Consider an instance \(I=(\varvec{p},\varvec{w},\varvec{r},m)\), and let \(I'=(\varvec{p}',\varvec{w},\varvec{r}',m)\), where \(\varvec{p}'\le \alpha \varvec{p}\) and \(\varvec{r}'\le \alpha \varvec{r}\). Then, we have
Proof
Let \(S_j^\Pi (I)\) denote the starting time of job j in schedule \(\Pi \) for instance I. We define a schedule \(\Pi \) with \(S_j^\Pi (I' )=\alpha \cdot C_j^{\text {OPT}}(I)- p_j'\). By definition \(C_j^\Pi (I')=\alpha \cdot C_j^{\text {OPT}}(I)\). Clearly, \(\Pi (I')=\alpha \cdot \text {OPT}(I)\). We claim that \(\Pi \) is feasible for \(I'\). Indeed, we have
Suppose now that two jobs j, k overlap in \(\Pi \), i.e., \(S_j^\Pi (I') < C_k^\Pi (I') \le C_j^\Pi (I')\). By definition of \(\Pi \), this means that \(\alpha \cdot C_j^{\text {OPT}}(I) - p_j' < \alpha \cdot C_k^{\text {OPT}}(I) \le \alpha \cdot C_j^{\text {OPT}}(I)\). Therefore, \(S_j^{\text {OPT}}(I) \le S_j^{\text {OPT}}(I) + p_j - \frac{p_j'}{\alpha } < C_k^{\text {OPT}}(I) \le C_j^{\text {OPT}}(I)\), meaning that j and k also overlap in the optimal schedule for I. Since the optimal schedule for I always schedules at most m jobs in parallel, this is also the case for \(\Pi \). Hence, \(\Pi \) is a feasible schedule for \(I'\) and \(\text {OPT}(I')\le \Pi (I')=\alpha \cdot \text {OPT}(I)\). \(\square \)
This lemma will turn out to be useful for both investigated settings.
6.1 Release dates
Let us first consider the single machine case where jobs arrive online. For instances of , we extend \({\mathfrak {D}_{b}}\) in the following way: The strategy keeps track of the rank \(q_j(\theta )\) of each job j at time \(\theta \), where \(q_j(\theta )\) is the largest integer q such that job j has already been probed for \(w_j b^{q-1}\) before \(\theta \). At any end of a probe occurring at time \(\theta \) it selects the job j with minimum rank and index among all released and not completed jobs and probes it for \(w_j b^{q_j(\theta )}\). We only consider the limit strategy obtained when the rank of a job is set to \(q_0\rightarrow -\infty \) at its release, so that phases of infinitesimal probing occur after each release date. Note that the strategy never interrupts a probe, i.e., each probe \((t_i, j_i, \tau _i)\) is executed until time \(t_i + \min \{p_j, \tau _i\}\). In order to show an upper bound on the competitive ratio of \({\mathfrak {D}_{b}}\), we actually compare \(\mathfrak {D}_{b}\) against an optimal preemptive offline algorithm. This means that we obtain an upper bound on the “power of preemption” in the online setting with restarts, complementing the unbounded ratio in the model without restarts.
Theorem 19
\({\mathfrak {D}_{b}}\) is \(\frac{2b^4}{2b^2-3b+1}\)-competitive for . This ratio is minimized for \(b=\frac{9+\sqrt{17}}{8}\), yielding a performance guarantee of \(\frac{107+51\sqrt{17}}{32} \approx 9.915\).
To prove this result, we need a bound on the end time of a probe with respect to the point in time at which the probe began.
Lemma 20
For \(I = (\varvec{p}, \varvec{w}, \varvec{r}, m)\) denote by S the schedule produced by \({\mathfrak {D}_{b}}\). If some job j is probed at time t for \(\tau = w_j b^{q}\) for some \(q\in \mathbb {Z}\), then \(t + \tau \le b t\).
Proof
At the start of the probe, \({\mathfrak {D}_{b}}\) has spent \(\sum _{i=-\infty }^{q-1} w_j b^{i} = w_j \frac{b^{q}}{b-1}\) time probing the job j. Since probes of the same job cannot run in parallel on multiple machines, we have that \(t \ge w_j \frac{b^{q}}{b-1}\). Thus,
\(\square \)
The proof of Theorem 19 consists of two main steps that are carried out in the subsequent two lemmas. In the first step, we compare an arbitrary instance I to an instance \(I'\) with processing times such that the Smith ratios are rounded to integer powers of b and release dates that are shifted to end times of probes. In the second step, we compare the performance of \({\mathfrak {D}_{b}}\) on the instance \(I'\) to the performance of \(\text {WSETF}\) on another instance \(I''\). Figure 2 illustrates these two auxiliary instances for an example.
An example for the three schedules considered in the proof of Theorem 19 for the instance I with \(\varvec{p} = (8,9,9,3,2,2)^{\top }\), \(\varvec{r} = (0,0,18,18,30,93)^{\top }\), and unit weights \(\varvec{w} = \varvec{1}\), for \(b = 2\). Gray areas indicate infinitesimal probing; thick lines indicate the completion of a job. Top: The schedule of \(\mathfrak {D}_2 (I)\) for the original instance I. Middle: The schedule of \(\mathfrak {D}_2 (I')\) for the modified instance \(I'\) with processing times rounded to the next integer power of \(b=2\). The release dates \(r_3'\) and \(r_4'\) are shifted such that they coincide with the end of a probing. Bottom: The schedule of \(\text {WSETF}\) for the instance \(I''\). The processing times (corresponding to the colored areas) correspond to the total probing times in the schedule of \(\mathfrak {D}_2 (I')\). The completion times in this schedule are higher or equal to the completion times in the schedule of \(\mathfrak {D}_2 (I')\), as indicated by the vertical dashed lines
Lemma 21
For an arbitrary instance \(I = (\varvec{p}, \varvec{w}, \varvec{r})\) of there exists another instance \(I' = (\varvec{p}', \varvec{w}, \varvec{r}')\) with \(\varvec{p}'\le \frac{b^3}{2b-1} \varvec{p}\) and \(\varvec{r}'\le \frac{b^3}{2b-1} \varvec{r}\) such that \({\mathfrak {D}_{b}}(I)\le {\mathfrak {D}_{b}}(I')\). Moreover, the instance \(I'\) has the property that for every job j there exists an integer \(q_j\in \mathbb {Z}\) with \(p_j'=w_j b^{q_j}\) and every release date either coincides with the end of a probe in the schedule \({\mathfrak {D}_{b}}(I')\) or is some point in time at which \({\mathfrak {D}_{b}}\) idles in \(I'\).
Proof
Let I be an arbitrary instance and consider the schedule S produced by \({\mathfrak {D}_{b}}\) for I. For a fixed job j with \(r_j>0\), we denote by \(\pi (j) = (t(j), k(j), \tau (j))\) the last probe that is started by \({\mathfrak {D}_{b}}\) before the release date of j, hence, \(\pi (j)\) satisfies \( t (j) < r_j \).
For every job j, let
where \(q_j {:}{=}\big \lceil \log _b \big ( \frac{p_j}{w_j} \big )\big \rceil \). Observe that \(p_j \le p_j' \le b p_j \le \frac{b^3}{2b-1} p_j\).
We define a new schedule \(\bar{S}\) as follows. For every probe \((t, k, \tau )\) in the schedule S, there is a corresponding probe \((t', k', \tau ')\) with \(k'=k\), \(\tau ' = \tau \) and \(t' \ge t\) in the schedule \(\bar{S}\), where the times \(t'\) are chosen such that no additional idle time exists in \(\bar{S}\). In particular, in the schedule \(\bar{S}\) the same jobs are probed for the same times in the same order as in the schedule S. However, the actual duration of any probe \((t', k', \tau ')\) in \(\bar{S}\) depends on the processing time \(p_j'\) rather than \(p_j\), i.e., the duration is \(\min \{p_j', \tau \}\). In particular, all probes have the same duration except for those, where jobs complete. These probes last longer and might shift all subsequent probes. We define new release dates \(\varvec{r}'\) as follows. For every job j with \(r_j > 0\), consider the probe \(\pi '(j) = (t'(j), k'(j), \tau '(j))\) in the schedule \(\bar{S}\) that corresponds to the probe \(\pi (j)\), and set \(r_j' {:}{=}\max \{ r_j, t' (j) + \tau '(j) \}\). For all jobs j with \(r_j = 0\), we set \(r_j' = 0\). Overall, we define a new instance
Then, the schedule produced by \({\mathfrak {D}_{b}}\) for \(I'\) is exactly \(\bar{S}\) by construction. In particular, we have \({\mathfrak {D}_{b}}(I) \le {\mathfrak {D}_{b}}(I')\).
We now argue that \(\varvec{r}' \le \frac{b^3}{2b-1} \varvec{r}\). To this end, consider a fixed job j. If \(r'_j=r_j\) there is nothing to show. Hence, let \(r'_j=t'+\tau '\). We compare the total time devoted to each job k before the start times of the probe performed when j is released. If k is not completed before t in \({\mathfrak {D}_{b}}(I)\), the total probing time of k before t in \({\mathfrak {D}_{b}}(I)\) is the same as the total probing time of k before \(t'\) in \({\mathfrak {D}_{b}}(I')\). If k is completed before t in \({\mathfrak {D}_{b}}(I)\), the total time devoted to job k is \(w_k\sum _{i=-\infty }^{q_k - 1} b^i + p_k = w_k \big ( \frac{b^{q_k}}{b-1} + \frac{p_k}{w_k}\big )\). On the other hand, in \({\mathfrak {D}_{b}}(I')\) the total time devoted to k is \(w_k\sum _{i=-\infty }^{q_k} b^i = w_k\frac{b^{q_k + 1}}{b-1}\). The ratio between these two elapsed times is
Since this holds for every job k, we have \(t' \le \frac{b^2}{2b-1} t\).
Using Lemma 20 we obtain
\(\square \)
For every time \(\theta > 0\) and job j, we denote by
the rank of job j at time \(\theta \) in \({\mathfrak {D}_{b}}(I')\), where \((t, j, w_j b^{q-1})\) describes the probe of job j starting at time t for a probing time of \(w_j b^{q-1}\) and \(I'\) is defined as in Lemma 21. In other words, \(q_j(\theta )=q\) holds if and only if job j has already been probed for \(w_j b^{q-1}\) before time \(\theta \), and the next probe of job j will be a probing for \(w_j b^q\) if not completed. Note that the function \(q_j (\theta )\) is piecewise constant and increases by 1 whenever a probe has ended. Observe that \(q_j (\infty )\) is exactly the value \(q_j\) defined in the proof of Lemma 21. Further, we denote by
the minimal rank among all jobs that are released and not completed at time \(\theta \). If there is no such job at time \(\theta \), we set \(q(\theta ) = \infty \). Note that the function \(q(\theta )\) decreases whenever new jobs are released and increases when the minimum rank of all released and not completed jobs increases. The definition is made so that q is left-continuous. We also define
Lemma 22
For an instance \(I' = (\varvec{p}', \varvec{w}, \varvec{r}')\) with the properties described in Lemma 21 the instance \(I'' = (\varvec{p}'', \varvec{w}, \varvec{r}')\) with \(\varvec{p}'' {:}{=}\frac{b}{b-1} \varvec{p}'\) satisfies the inequality \({\mathfrak {D}_{b}}(I') \le \text {WSETF}(I'')\).
Proof
Observe that
i.e., \(p''_j\) is obtained by adding all failing times of j to the processing time \(p'_j\). We show for all jobs j that
Let \(0 = r^{(0)}< \cdots < r^{(n')}\) be all distinct release dates, and for every \(i \in \{0,1,\dotsc ,n'\}\) and \(q\in \mathbb {Z}\) let \(J^{(i)} {:}{=}\{j \in [n] \mid r_j = r^{(i)}\}\) and
be the end time of round \(q\) of jobs released at \(r^{(i)}\), i.e., at time \(e^{(i)}(q)\) every job \(j \in J^{(i)}\) not completed at an earlier round has been probed for \(\sum _{\widehat{q}= -\infty }^qw_jb^{\widehat{q}}=w_j\frac{b^{q+1}}{b-1}\) in total in \({\mathfrak {D}_{b}}(I')\).
In the following we will show that for all \(q\in \{q_{\min }- 1, q_{\min }, \dotsc ,q_{\max }\}\) and for all \(i \in \{0,\dotsc ,n'\}\) every job \(j\in J^{(i)}\) has the same elapsed time at \(e^{(i)}(q)\) in the schedule computed by \({\mathfrak {D}_{b}}\) for \(I'\) and the schedule computed by \(\text {WSETF}\) for \(I''\), i.e., \(Y^{{\mathfrak {D}_{b}}}_j(I', e^{(i)}(q)) = Y^{\text {WSETF}}_j(I'', e^{(i)}(q))\). This will be done by induction on \(q\). Figure 3 illustrates the end times of rounds \(e^{(i)}(q)\) as well as the value \(q_{\min }\). For the sake of simplicity, we write for the remainder of the proof \(Y^{{\mathfrak {D}_{b}}}_j (t) = Y^{{\mathfrak {D}_{b}}}_j (I',t)\) and \( Y^{\text {WSETF}}_j (t) = Y^{\text {WSETF}}_j (I'',t)\). We start with the base case \(q= q_{\min }-1\), i.e., we show that for all \(i\in \{0,\ldots ,n'\}\) and for all \(j\in J^{(i)}\) we have

Illustration of the situation in the proof of Lemma 22. The release dates \(r^{(i)}\) and the end points of the rounds in \({\mathfrak {D}_{b}}\) subdivide the time axis into intervals. The shaded intervals do not contain any successful probings. The value \(q_{\min }= 5\) is chosen such that the end of the round \(e^{(i)}(q_{\min }- 1)\) is before the next release date and such that between \(r^{(i)}\) and \(e^{(i)}(q_{\min }- 1)\) no job completes. Equation (19) refers to the endpoints \(e^{(i)}(q_{\min }- 1)\), while Equation (23) refers to all subsequent endpoints of rounds. Note, that some endpoints (e.g, \(e^{(2)}(6)\) or \(e^{(2)} (7)\)) lie after one or more subsequent release dates
For every i, the definition of \(e^{(i)}\) and the left-continuity of q imply that \(q(t) < q_{\min }\) for all \(t \in (r^{(i)}, e^{(i)}(q_{\min }-1)]\). Therefore, no job is released or completed by \({\mathfrak {D}_{b}}\) during \((r^{(i)},e^{(i)}(q_{\min }-1)]\). Hence,
Assume for a contradiction that Equation (19) is wrong for some i and j, and let i be the minimum index for which the equation is violated for some job j. Clearly, at time \(r^{(i)}\) jobs from \(J^{(i)}\) have neither been processed in \(\text {WSETF}\) nor in \({\mathfrak {D}_{b}}\), while all earlier released jobs with \(r_j < r^{(i)}\) have either been completed in both \(\text {WSETF}\) and \({\mathfrak {D}_{b}}\) or have in both schedules an elapsed time of at least \(w_j \frac{b^{q_{\min }}}{b-1}\), using (20) and the minimality assumption. Thus, at time \(r^{(i)}\), both schedules start by processing only jobs from \(J^{(i)}\). For \({\mathfrak {D}_{b}}\) we know that these are contiguously processed at least until time \(e^{(i)}(q_{\min }-1)\). Since no job from \(J^{(i)}\) is completed before \(e^{(i)}(q_{\min }- 1)\), we know that \(p_j' \ge w_j b^{q_{\min }}\), and thus \(p_j'' \ge w_j \frac{b^{q_{\min }+1}}{b-1}\) for all \(j \in J^{(i)}\). Moreover, for any job \(j\in J^{(i)}\) we have
Therefore,
Until the first completion time \(\text {WSETF}\) processes each job \(j \in J^{(i)}\) with a rate of \(w_j/\sum _{k \in J^{(i)}} w_k\). Suppose the first completing job j is completed at time \( t < e^{(i)}(q_{\min }-1)\). Then it has elapsed time
a contradiction since j would have elapsed time strictly less than its processing time. Therefore, we have \(C_j^{\text {WSETF}}(I'') \ge e^{(i)}(q_{\min }-1)\) for all \(j \in J^{(i)}\), and the elapsed times of these jobs j at this time can be computed as
This contradicts the assumption that Equation (19) is wrong for i.
Now we show an equation analogous to Equation (19) for arbitrary \(q \ge q_{\min }\), i.e., we show that for all \(q \in \{q_{\min },\dotsc ,q_{\max }\}\), for all \(i \in \{0,\dotsc ,n'\}\), and for all \(j \in J^{(i)}\) we have
Moreover, we will show that if \(q_j \ge q\), then this quantity is equal to \(w_j\frac{b^{q+1}}{b-1}\), and if \(q_j = q\), then \(C_j^{\text {WSETF}}(I'')=e^{(i)}(q)\), where we are using the notation from Lemma 21, so that \(p_j' = b^{q_j}\) for every \(j \in [n]\), and thus \(p_j'' = \frac{b^{q_j+1}}{b-1}\).
As announced, this is shown by induction on q. For \(q = q_{\min }-1\), the statements were shown in Equation (19) because every job has \(q_j > q_{\min }- 1\). Hence, in the inductive proof we can assume for \(q \ge q_{\min }\) that the claim is true for all \(q'\) with \(q_{\min }- 1 \le q' < q\). For all \(i \in \{0,\dotsc ,n'\}\), all jobs \(j \in J^{(i)}\) with \(q_j < q\) are completed in \({\mathfrak {D}_{b}}\) before time \(e^{(i)}(q_j)\). The induction hypothesis implies that \(Y_j^{{\mathfrak {D}_{b}}}(e^{(i)}(q_j)) = Y_j^{\text {WSETF}}(e^{(i)}(q_j)) = w_j \frac{b^{q_j + 1}}{b-1} = p_j''\). Hence, also in the \(\text {WSETF}\) these jobs are completed until time \(e^{(i)}(q_j) \le e^{(i)}(q-1)\). In the following we thus restrict to jobs with \(q_j \ge q\).
Assume for a contradiction that the claim is wrong for q, and let i be the smallest index for which the statement fails. Let \(e {:}{=}e^{(i)}(q)\), and let \(\mathcal I(e) {:}{=}\bigl \{i' \in \{0,\dotsc ,n'\} \mid e^{(i')}(q) = e\bigr \} = \{\underline{i},\underline{i}+1,\dotsc ,\overline{i}\}\). Let
All not completed jobs \(j \in J^{(i')}\) with \(i' < \underline{i} \le i\) have \(e^{(i')}(q) < r^{(\underline{i})}\), and hence, as i was chosen to be minimal, elapsed time of at least \(w_j\frac{b^{q+1}}{b-1}\) in the schedules constructed by both \({\mathfrak {D}_{b}}\) and \(\text {WSETF}\). Therefore, none of these jobs are probed in the interval \([r^{(\underline{i})}, e]\) by \({\mathfrak {D}_{b}}\). We know by the property of \(I'\) that release dates are only at ends of probes, which implies that during the union of intervals
each job \(j \in J'\) is probed for \(w_j b^q\) by \(\mathfrak {D}_{b}\). Observe that in this formula it is possible that the left bound of an interval is not smaller than the right bound in which case the interval is empty. Therefore, using the property of \(I'\) that the duration of each probe is exactly the probing time itself we have
where \((\cdot )^+\) denotes the positive part. We next argue that in the \(\text {WSETF}\) schedule, for all \(t \in \mathcal T \setminus \{e\}\) every job \(j \in J'\) has elapsed time \(Y_j^{\text {WSETF}}(t) < w_j \frac{b^{q+1}}{b-1}\). Since all \(q_j \ge q\) for \(j \in J'\), this, along with the lower bound for the elapsed processing times of previously released jobs shown above, implies that also \(\text {WSETF}\) is not running any jobs released before \(r^{(\underline{i})}\) during this time, and, in view of the construction of the instance \(I''\), that no job from \(J'\) is completed within this time. To show the claimed statement, consider the first moment in time \(t \in \mathcal T\) at which some job \(j \in J'\) has elapsed processing time \(Y_j^{\text {WSETF}}(t) \ge w_j \frac{b^{q+1}}{b-1}\). Let \(i^* \in \mathcal I(e)\) be the index such that \(t \in (e^{(i^*)}(q-1), r^{(i^*+1)})\). Since \(\text {WSETF}\) processes j directly before t, the job j must have minimum weighted elapsed time among all released and unfinished jobs, so in particular among all jobs \(k \in \bigcup _{i'=\underline{i}}^{i^*} J^{(i')}\) with \(q_{k} \ge q\), i.e.,
Furthermore, no job \(k \in J' \cap J^{(i')}\) for \(i' \in \mathcal I(e)\) is processed by \(\text {WSETF}\) during an interval \((r^{(i'')}, e^{(i''+1)}(q-1)]\) for some \(i'' > i'\), so that the time spent on k during \(\mathcal T \cap (0,t)\) is simply \(Y_k^{\text {WSETF}}(t) - Y_k^{\text {WSETF}}(e^{(i')}(q-1))\). Therefore, \(\text {WSETF}\) has in total spent
on jobs \(k \in J' \cap \bigcup _{i'=\underline{i}}^{i^*} J^{(i')}\) during \(\mathcal T \cap (0,t)\), where we used the induction hypothesis for q. Thus, within \(\mathcal T \cap (0,t)\), the strategy \({\mathfrak {D}_{b}}\) has had enough time to probe all jobs from this set for \(w_k b^{q}\). In particular it has probed the jobs from \(J^{(\underline{i})}\) for this time, so by definition of \(e^{(\underline{i})}(q)\), we have \(t \ge e^{(\underline{i})} (q) = e\), hence \(t = e\).
By continuity, all jobs \(j \in J'\) have \(Y_j^{\text {WSETF}}(e) \le w_j \frac{b^{q+1}}{b-1}\). As \(\mathcal D_b\) manages to process all these jobs long enough during \(\mathcal T\) so that at time e they have accumulated exactly this amount of elapsed time and \(\text {WSETF}\) always processes jobs from \(J'\) during \(\mathcal T\), it must have assigned in total the same amount of time within \(\mathcal T\) to jobs from \(J'\), i.e., \(\sum _{j \in J'} Y_j^{\text {WSETF}}(e) = \frac{b^{q+1}}{b-1} \sum _{j \in J'} w_j\). Together with the above upper bound for each individual job \(j \in J'\), this implies that \(Y_j^{\text {WSETF}}(e) = w_j \frac{b^{q+1}}{b-1}\) for all \(j \in J'\). Consequently, \(C_j^{\text {WSETF}}(I'') = e\) if \(q_j = q\). Thus, the statement claimed above is true for all \(i' \in \mathcal I(e)\), so in particular for i, contradicting our assumption.
In total, we have thus shown that \(C^{{\mathfrak {D}_{b}}}_j (I') \le e^{(i)}(q_j) = C^{\text {WSETF}}_j (I'')\) for all jobs \(j \in J^{(i)}\) for all i, and thus, \({\mathfrak {D}_{b}}(I') \le \text {WSETF}(I'')\). \(\square \)
Proof of Theorem 19
By Theorem 12 we know that \(\text {WSETF}\) is 2-competitive for . Moreover, using Lemmas 21 and 22 together with Lemma 18, we obtain
\(\square \)
6.2 Parallel machines
In this subsection we consider instances \((\varvec{p}, \varvec{1}, \varvec{0}, m)\) of with processing times \(\varvec{p}\), unit weights, trivial release dates, and m identical parallel machines. For this setting we have to extend the definition of kill-and-restart strategies and in particular of the b-scaling strategy. The set of all currently active probes are added to each state. The intervals chosen by an action need not be disjoint anymore, but instead it is required that any point in time t be covered by at most m intervals. Finally, the transition function \(T_I\), mapping a state and an action to a new state, requires several modifications: It has to deal with the situation that multiple jobs are simultaneously completed, the elapsed probing times of the active probes given in the state have to be taken into account in the determination of the next completion time, and the active probes at the next decision time have to be determined. We do not go into more detail for the general definition.
For the b-scaling strategy \({\mathfrak {D}_{b}}\), we perform the same sequence of probes, but assign them to the parallel machines in a list scheduling manner, i.e., every probe is scheduled on the first available machine. Moreover, at the moment when the number of remaining jobs becomes less than or equal to the number of machines, the jobs are not aborted anymore. The formalization of the actions chosen upon the completion of any job is straightforward. However, it may be not obvious how to formalize the action chosen at time 0 if n is not divisible by m, because then the last planned probes of each round q need not be synchronized. The infinitesimal probing makes it impossible to define the probing intervals in an inductive way. This can be resolved by observing that the planned probes are scheduled in an SPT manner, and it is well-known that in an SPT schedule, every mth job goes to the same machine. Hence, it is possible to split the family of all probe operations into m subsequences, each containing every mth element. More precisely, if the probes of the single-machine strategy are denoted by \(\pi _{(q,j)} = (t_{(q,j)}, j, \tau _{(q,j)})\) for \((q,j) \in \mathbb {Z}\times [n]\), then consider the bijection \(\iota :\mathbb {Z}\times [n] \rightarrow \mathbb {Z}\) with \(\iota (q,j) {:}{=}nq + j\), specifying the probing order, and the subfamilies \(a_i = (\pi _{\iota ^{-1}(k m + i)})_{k \in \mathbb {Z}}\) of probes to be assigned to each machine \(i \in [m]\). We define the probes of the initial action for m machines as
where \(\sum _{k=-\infty }^{k-1} \tau _{\iota ^{-1}(km+i)} < \infty \) because of the absolute convergence of the sequence of all probing times smaller than any given one. The entire initial action of the m-machine b-scaling strategy is then the family \((\pi '_{(i,k)})_{i \in [m], k \in \mathbb {Z}}\).
Theorem 23
\({\mathfrak {D}_{b}}\) is \(\frac{3b^2-b}{b-1}\)-competitive for . This ratio is minimized for \(b=\frac{3+\sqrt{6}}{3}\), yielding a performance guarantee of \(5+2\sqrt{6} \approx 9.899\).
Proof
Let \(I = (\varvec{p}, \varvec{1}, \varvec{0}, m)\) be an arbitrary instance for . We assume that \(n > m\) as otherwise \({\mathfrak {D}_{b}}\) is optimal. We define a new instance \(I' = (\varvec{p}', \varvec{1}, \varvec{0}, m)\) with \(p_j':=b^{q_j}\) and \(q_j = \lceil \log _b(p_j) \rceil \). By definition of \({\mathfrak {D}_{b}}\), the last job executed last on each machine is run non-preemptively, i.e., it is probed for an infinite amount of time. We denote by \(\widehat{J}\) this set of m jobs run non-preemptively in instance \(I'\).
We first show that \({\mathfrak {D}_{b}}(I)\le {\mathfrak {D}_{b}}(I')\). Denote by \(\pi _0=(t_0,j_0,\tau _0)\) the first probe in \({\mathfrak {D}_{b}}(I)\) in which a job is completed and denote by \(\pi _0,\pi _1,\ldots ,\pi _N\) the sequence of all probes started at or after time \(t_0\), ordered by starting time of the probing operation. Each probe \(\pi _k=(t_k,j_k,\tau _k)\) is in one-to-one correspondence with a probe \(\pi _k'=(t_k',j_k,\tau _k)\) in \({\mathfrak {D}_{b}}(I')\). Denote by \(\lambda _{k1}\le \lambda _{k2} \le \ldots \le \lambda _{km}\) the ordered loads of the m machines in \({\mathfrak {D}_{b}}(I)\) at time \(t_k\), where the load of a machine at time \(\theta \) is the last end of a probe started before \(\theta \) on that machine. In particular, \(\lambda _{k1}=t_k\) because probe \(\pi _k\) starts at time \(t_k\) on the least loaded machine. Similarly, denote by \(\lambda _{k1}' \le \ldots \le \lambda _{km}'\) the m ordered machine loads at time \(t_k'\) in \({\mathfrak {D}_{b}}(I')\).
We show by induction on k that \(\lambda _{ki}\le \lambda _{ki}'\) holds for all \(i\in [m]\). The base case \(k=0\) is trivial, since the probes are identical in \({\mathfrak {D}_{b}}(I)\) and \({\mathfrak {D}_{b}}(I')\) until time \(t_0\), hence \(\lambda _{0i}=\lambda _{0i}', \forall i\in [m]\). Then, let \(k\in \{0,\ldots ,N-1\}\), and denote by \(\delta _k=\min (p_{j_k},\tau _k)\) the actual duration of the probing operation \(\pi _k\), so the loads \(\varvec{\lambda }_{k+1}\) at time \(t_{k+1}\) in \({\mathfrak {D}_{b}}(I)\) are a permutation of \((\lambda _{k1}+\delta _k,\lambda _{k2},\ldots ,\lambda _{km})\). Similarly, the loads \(\varvec{\lambda }_{k+1}'\) at time \(t_{k+1}'\) in \({\mathfrak {D}_{b}}(I')\) are a permutation of \((\lambda _{k1}'+\delta _k',\lambda _{k2}',\ldots ,\lambda _{km}')\), with \(\delta _k'=\min (p_{j_k}',\tau _k)\ge \delta _k\). By induction hypothesis we have \(\lambda _{k1}+\delta \le \lambda _{k1}'+\delta \le \lambda _{k1}'+\delta '\) and \(\lambda _{ki}\le \lambda _{ki}'\) for all \(i=2,\ldots ,m\). This shows the existence of two permutations \(\sigma \) and \(\sigma '\) such that \(\lambda _{k+1,\sigma (i)}\le \lambda _{k+1,\sigma '(i)}'\) holds for all \(i\in [m]\), which in turn implies that the ordered loads satisfy \(\lambda _{k+1,i}\le \lambda _{k+1,i}'\), for all \(i\in [m]\). This concludes the induction. The end of probe \(\pi _k\) in \({\mathfrak {D}_{b}}(I)\) is \(t_k+\delta _k=\lambda _{1k}+\delta _k\le \lambda _{1k}'+\delta _k'=t_k'+\delta _k'\), where the latter corresponds to the end of probe \(\pi _k'\) in \({\mathfrak {D}_{b}}(I')\). Clearly, this implies \({\mathfrak {D}_{b}}(I)\le {\mathfrak {D}_{b}}(I')\).
Let \(q_{\min }{:}{=}\min _j q_j\) and \(q_{\max }{:}{=}\max _{j \notin \widehat{J}} q_j\). For \(q<q_{\max }\) denote by \(T'_i (q)\) the last end of a \(b^q\)-probing operation on machine i in \({\mathfrak {D}_{b}}(I')\). For \(q=q_{\max }\) we need to define \(T_i'(q_{\max })\) differently to take into account the m jobs probed for an infinite time. Note that the jobs \(j\in \widehat{J}\) have \(p_j'\ge b^{q_{\max }}\) and for each \(i\in [m]\) there is exactly one job \(j(i)\in \widehat{J}\) that is completed on machine i. If job j(i) has already been probed for \(b^{q_{\max }}\) before being run non-preemptively, we define \(T_i'(q_{\max })\) as the last end of a \(b^{q_{\max }}\)-probing operation on i; otherwise we define \(T_i'(q_{\max })\) as the first point in time where j(i) has been processed for at least \(b^{q_{\max }}\). This ensures that every job with \(p_j'\ge b^{q_{\max }}\) is processed on some machine \(i\in [m]\) during one interval of length \(b^{q_{\max }}\) contained in \([T_i'(q_{\max }-1),T_i'(q_{\max })]\).
We define another instance \(I'' = (\varvec{p}'', \varvec{1}, \varvec{0}, m)\) with processing times
For all j such that \(q_j\le q_{\max }\) we have \(p_j'' = \frac{b^{q_j+1}}{b-1} = \frac{b}{b-1} p_j'\) and otherwise we have \(p_j'' = \frac{b^{q_{\max }+1}}{b-1} + p_j' \le \bigl ( \frac{1}{b-1} + 1 \bigr ) p_j' = \frac{b}{b-1} p_j'\), i.e., \(p''_j \le \frac{b}{b-1} p'_j\) holds for all j.
Consider the schedule \(S''\) of \(\text {RR}\) for the instance \(I''\). Denote by \(T''(q)\) the first point in time t with \(Y_j^{\text {RR}} (t) = \frac{b^{q+ 1}}{b-1}\) for all jobs j with \(p_j'' \ge \frac{b^{q+ 1}}{b-1}\). In particular, every job with \(q_j \le q_{\max }\) is completed at \(T''(q_j)\) in the schedule \(S''\). We first prove by induction on \(q\) that \(\sum _{i=1}^m T'_i (q) = m \, T''(q)\), for all \(q=q_{\min },\ldots ,q_{\max }\). Since for all jobs j we have \(q_j \ge q_{\min }\) no job is completed in \({\mathfrak {D}_{b}}\) and \(\text {RR}\) until \(T'_i (q_{\min }-1)\) for any machine i and \(T''(q_{\min }-1)\), respectively. Thus, we have
Let \(q> q_{\min }\). At \(T''(q-1)\), there are, by definition, more than m remaining jobs and each job j with \(q_j \ge q\) has already been processed for \(\frac{b^{q+ 1}}{b-1}\) in the \(\text {RR}\)-schedule for instance \(I''\). Therefore, each job receives a rate of \(\frac{m}{n_{\ge q}}<1\) in this round, where \(n_{\ge q}\) is the number of jobs with \(p'' \ge \frac{b^{q+ 1}}{b-1}\) and, thus, \(T''(q)=T''(q-1)+\frac{n_{\ge q}}{m}\cdot b^{q}\). On the other hand, if \(q<q_{\max }\) \({\mathfrak {D}_{b}}\) probes \(n_{\ge q}\) jobs for exactly \(b^{q}\), so \(\sum _{i=1}^m T'_i (q) = \sum _{i=1}^m T'_i(q-1) + n_{\ge q} b^{q}\) because there is no idling-time in the schedule. For \(q=q_{\max }\), our definition of \(T_i'(q_{\max })\) ensures that \(\sum _{i=1}^m T'_i (q_{\max }) = \sum _{i=1}^m T'_i(q_{\max }-1) + n_{\ge q_{\max }} b^{q_{\max }}\) holds as well. Then, the claim follows from the induction hypothesis.
Consider a job \(j \notin \widehat{J}\). We have
where the second inequality comes from the fact that the probing operations are done in a list-scheduling manner. For a job \(j \in \widehat{J}\), we have \(C^{{\mathfrak {D}_{b}}}_j (I') \le T'_{i(j)} (q_{\max }) +p_j'\), where i(j) denotes the machine on which j is probed for an infinite amount of time. This implies
Since \(\text {RR}\) is 2-competitive for (see [5]), we overall obtain
where we used Lemma 18 for the last two inequalities. \(\square \)
7 Conclusion
We studied kill-and-restart as well as preemptive strategies for the problem of minimizing the sum of weighted completion times and gave a tight analysis of the deterministic and randomized version of the natural b-scaling strategy for as well as of \(\text {WSETF}\) for
.
We hope that this work might lay a basis for obtaining tight bounds on the performance of the b-scaling strategy for more general settings such as non-trivial release dates and parallel machines. Moreover, we think that the class of kill-and-restart strategies combines the best of two worlds. On the one hand, they allow for interruptions leading to small competitive ratios in contrast to non-preemptive algorithms, on the other hand, they reflect the non-preemptive property of only completing a job if it has been processed as a whole.
8 The b-scaling strategy with infinitesimal probing
In this section, we briefly discuss how to define the b-scaling strategy with infinitesimal probing. The b-scaling strategy \({\mathfrak {S}_{b}}^{\text {id},0,q_{0}}\) without infinitesimal probing starts at a fixed round \(q_0\) and is given in Algorithm 1. While \({\mathfrak {S}_{b}}^{\text {id},0,q_{0}}\) can easily be implemented, it is not possible to implement the limit strategy \({\mathfrak {S}_{b}}^{\text {id},0,-\infty }\), for example, on a Turing machine, since at a time arbitrarily close to 0 it has probed each jobs an infinite number of times.
Let us now formally define the strategy \({\mathfrak {S}_{b}}^{\sigma ,\xi , -\infty }\), by describing the action \(a(s)=\big ((t_i,j_i,\tau _i)\big )_{i\in \mathcal {I}}\in \mathcal {A}(s)\) it takes in any state \(s=(\theta , U, \varvec{\mu })\), in accordance with the kill-and-restart framework described in Sect. 2. Recall that an action is a family of planned probes that the strategy is committed to execute until a job completes and a new action is determined. Moreover, the state s specifies lower bounds \(\mu _j\le p_j\) for every job j, the set U of unfinished jobs, and the current time \(\theta \).
In the initial state \(s_0=(0, [n], \varvec{0})\), we plan to probe all jobs \(j\in [n]\) in rounds, where in each round the jobs are probed in the order given by \(\sigma \) for \(w_jb^{q+\xi }\) for some \(q\in \mathbb {Z}\) and then q incremented by 1. Hence, \(\sum _{\widehat{q}=-\infty }^{q-1}\sum _{k\in [n]}w_kb^{\widehat{q}+\xi }=\frac{b^{q+ \xi }}{b-1} \sum _{k\in [n]} w_k\) is the point in time at which the first job \(j =\sigma ^{-1}(1)\) is probed for \(w_jb^{q+\xi }\). We define the action of \({\mathfrak {S}_{b}}^{\sigma ,\xi }\) for state \(s_0\) by
In a state \(s=(\theta , U, \varvec{\mu })\) with \(\theta > 0\) occurring at the completion of a job, there exists \(q^*\in \mathbb {Z}\) by construction such that \(\mu _j\in \{w_j b^{q^*+\xi -1}, w_j b^{q^*+\xi }\}\), for all \(j\in U\). The set of jobs \(J^*{:}{=}\{j\in U \mid \mu _j=w_jb^{q^*+\xi -1}\}\) are those jobs that have not been probed yet for \(w_jb^{q^*+\xi }\). Hence, these jobs must be probed first before the new round \(q^*+1\) can start. For \(q>q^*\) we define
as the point in time when round q starts. We define the actions of \({\mathfrak {S}_{b}}^{\sigma ,\xi }\) for state s by
9 Technical lemmas
Lemma 24
Let \(L \in \mathbb {N}_{<0}\), \(b \ge 1\), and \(\varvec{B} = (\frac{1}{2} b^{\min (\ell , m)})_{0 \le \ell , m \le L}\) Then the Cholesky decomposition of \(\varvec{B}\) is \(\varvec{B} = \varvec{Y}^\top \varvec{Y}\) with \(\varvec{Y} = \bigl (\sqrt{\frac{b^\ell - b^{\ell -1} \cdot \mathbb {1}_{\ell \ge 1}}{2}} \cdot \mathbb {1}_{m \ge \ell }\bigr )_{0 \le \ell , m \le L}\).
Proof
Obviously,
is an upper triangular matrix with positive diagonal elements. An easy computation shows that \(\varvec{Y}^{\top } \varvec{Y} = \varvec{B}\). \(\square \)
Lemma 25
Let \(L \in \mathbb {N}_{>0}\), \(b \ge 1\), and \(0 = a_0 < a_1 \le \cdots \le a_L\). Let \(\varvec{A} = (\frac{1}{2} a_{\vert m-\ell \vert } \cdot b^{\min (\ell ,m)})_{0 \le \ell ,m \le L}\), and \(\varvec{B} = (\frac{1}{2} b^{\min (\ell ,m)})_{0 \le \ell ,m \le L}\). Then
where \(\varvec{Z} = (Z_{\ell ,m})_{0\le \ell ,m \le L}\) with
Proof
Let \(\varvec{B} = \varvec{Y}^\top \varvec{Y}\) be the Cholesky decomposition of \(\varvec{B}\). We can rewrite
So it remains to compute the matrix \(\varvec{Z}\). The matrix \(\varvec{Y}\) is given in Lemma 24, and its inverse is given by
By computing the product \(\varvec{Z} = \varvec{Y}^{-\top } \varvec{A} \varvec{Y}^{-1}\), we see that \(\varvec{Z}\) has the form claimed in the lemma, which is
\(\square \)
Lemma 26
Let \(b \ge 1\). For \(L \in \mathbb {N}_{>0}\) define \(\varvec{Y}_L {:}{=}\bigl (\sqrt{\frac{b^\ell - b^{\ell -1} \cdot \mathbb {1}_{\ell \ge 1}}{2}} \cdot \mathbb {1}_{m \ge \ell }\bigr )_{0 \le \ell , m \le L}\), \(\varvec{z}_L {:}{=}\bigl (\sqrt{\frac{2}{L+1}} \cdot \sin \bigl (\frac{\ell \pi }{L+1}\bigr )\bigr )_{0 \le \ell \le L}\), and \(\varvec{x}_L = (x_\ell ^{(L)}) {:}{=}\varvec{Y}^{-1} \varvec{z}_L\). Then \(\vert x_\ell ^{(L)}\vert \le \frac{2(\sqrt{b} + 1)}{\sqrt{(L+1)b^\ell (b-1)}}\) for all \(0 \le \ell \le L\), and there is an \(\ell ^* \in \mathbb {N}_{>0}\) such that for all \(L \ge \ell \ge \ell ^*\) we have that \(x^{(L)}_\ell \ge 0\). Moreover, \(\lim _{L \rightarrow \infty } \sum _{\ell =\ell ^*}^L x_\ell ^{(L)} = 0\).
Proof
Set \(\ell ' {:}{=}\bigl \lceil \frac{2}{\sqrt{b} - 1}\bigr \rceil \). We bound
so there is an \(L^*\) such that for all \(L \ge L^*\) the left-hand side is bounded by \(\sqrt{b}\). Set \(\ell ^* {:}{=}\max \{\ell ', L^*\}\), and let \(L \ge \ell ^*\) be fixed. By computing the product of the matrix in Equation (25) with \(\varvec{z}_L\) we obtain
This implies the bound on the absolute values. Since the function
is decreasing on [1, L], for all \(\ell \in \{\ell ^*,\dotsc ,L\}\) it holds that
To prove the last claim, observe that \(\sum _{\ell =\ell ^*}^L x^{(L)}_\ell \) is a telescoping sum for every \(L \ge \ell ^*\), and hence
\(\square \)
Lemma 27
For \(k, L \in \mathbb {N}\) denote by \(\varvec{T}_{k.L}\) the \(L \times L\) Toeplitz matrix with 2 on the main diagonal and \(-1\) on the kth and the \((-k)\)th superdiagonal. Let \(k, L \in \mathbb {N}\) with \(k \mid L\), let \(\varvec{v} \in \mathbb {R}^k\), and let \(\alpha \ge \Vert \varvec{v} \Vert ^2\). Then the matrix

is positive semidefinite.
Proof
We use the Schur complement lemma to show that \(\varvec{H}_L(\alpha , \varvec{v})\) is positive semidefinite. Let \(a {:}{=}L/k \in \mathbb {N}\), and observe that the matrix \(\varvec{T}_{k,L}\) is of the form \(\varvec{T}_{1,a} \otimes \varvec{I}_k\), where \(\otimes \) denotes the Kronecker product. The matrix \(\varvec{T}_{1,a}\) is a symmetric tridiagonal Toeplitz matrix, which has minimum eigenvalue \(\lambda _{\min }(\varvec{T}_{1,a}) = 2 \bigl (1-\cos \bigl (\frac{\pi }{a+1}\bigr )\bigr ) > 0\) (see [51, Theorem 2.4]), and is thus positive definite. The reader may verify that the inverse is given by \((\varvec{T}_{1,a}^{-1})_{ij}=\frac{1}{a+1} \cdot \min (i,j) \cdot (a+1-\max (i,j))\). Since the eigenvalues of the Kronecker product are the products of the eigenvalues, we have \(\lambda _{\min }(\varvec{T}_{k,L}) = \lambda _{\min }(\varvec{T}_{1,a}) > 0\), i.e., \(\varvec{T}_{k,L} \succ 0\), and, moreover, \(\varvec{T}_{k,L}^{-1}= \varvec{T}_{1,a}^{-1} \otimes \varvec{I}_k\). In particular, the upper left \(k\times k\) block of \(\varvec{T}_{k,L}^{-1}\) is equal to \((\varvec{T}_{1,a})^{-1}_{11} \cdot \varvec{I}_k=\frac{a}{a+1}\varvec{I}_k\). So we can form the Schur complement
This concludes the proof that \(\varvec{H}_L(\alpha ,\varvec{v})\) is positive semidefinite. \(\square \)
Lemma 28
Let \(a_{mn} \in \mathbb {R}\) for all \(m, n \in \mathbb {N}\). Assume that for every \(n \in \mathbb {N}\) the sequence \((a_{mn})_{m \in \mathbb {N}}\) converges to some \(a_n \in \mathbb {R}\) and that the sequence \((a_n)_{n \in \mathbb {N}}\) converges to some \(a \in \mathbb {R}\). Then there is \(s :\mathbb {N}\rightarrow \mathbb {N}\) so that \(a_{s(n)n} \xrightarrow {n \rightarrow \infty } a\).
Proof
For every \(n \in \mathbb {N}\) there is an \(s(n) \in \mathbb {N}\) such that \(\vert a_{s(n)n} - a_n \vert < \frac{1}{n}\). Then the resulting sequence \((a_{s(n)n})_{n \in \mathbb {N}}\) converges to a because for every \(\varepsilon > 0\) there is an \(N \in \mathbb {N}\) such that \(\vert a_n - a \vert < \frac{\varepsilon }{2}\) and \(\frac{1}{n} < \frac{\varepsilon }{2}\) for all \(n \ge N\), and hence, \(\vert a_{s(n)n} - a \vert \le \vert a_{s(n)n} - a_n \vert + \vert a_n - a \vert< \frac{1}{n} + \frac{\varepsilon }{2} < \varepsilon \). \(\square \)
Availability of data and materials
Not applicable.
References
Jäger, S., Sagnol, G., Schmidt genannt Waldschmidt, D., Warode, P.: Competitive kill-and-restart and preemptive strategies for non-clairvoyant scheduling. In: Del Pia, A., Kaibel, V. (eds.) Integer Program. Combin. Optim. (IPCO). LNCS, vol. 13904, pp. 246–260. Springer, Cham (2023). https://doi.org/10.1007/978-3-031-32726-1_18
Smith, W.E.: Various optimizers for single-stage production. Nav. Res. Logist. Q. 3(1–2), 59–66 (1956). https://doi.org/10.1002/nav.3800030106
Feldmann, A., Sgall, J., Teng, S.-H.: Dynamic scheduling on parallel machines. Theor. Comput. Sci. 130(1), 49–72 (1994). https://doi.org/10.1016/0304-3975(94)90152-X
Shmoys, D.B., Wein, J., Williamson, D.P.: Scheduling parallel machines on-line. SIAM J. Comput. 24(6), 1313–1331 (1995). https://doi.org/10.1137/S0097539793248317
Motwani, R., Phillips, S., Torng, E.: Nonclairvoyant scheduling. Theor. Comput. Sci. 130(1), 17–47 (1994). https://doi.org/10.1016/0304-3975(94)90151-1
Kim, J., Chwa, K.: Non-clairvoyant scheduling for weighted flow time. Inf. Process. Lett. 87(1), 31–37 (2003). https://doi.org/10.1016/S0020-0190(03)00231-X
Beaumont, O., Bonichon, N., Eyraud-Dubois, L., Marchal, L.: Minimizing weighted mean completion time for malleable tasks scheduling. In: Oliker, L., Yelick, K. (eds.) 26th IEEE International Parallel and Distributed Processing Symposium (IPDPS), pp. 273–284 (2012). https://doi.org/10.1109/ipdps.2012.34
Im, S., Kulkarni, J., Munagala, K.: Competitive algorithms from competitive equilibria: non-clairvoyant scheduling under polyhedral constraints. J. ACM 65(1), 1–33 (2017). https://doi.org/10.1145/3136754
Im, S., Kulkarni, J., Munagala, K., Pruhs, K.: SelfishMigrate: A scalable algorithm for non-clairvoyantly scheduling heterogeneous processors. In: Barak, B. (ed.) 55th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pp. 531–540 (2014). https://doi.org/10.1109/FOCS.2014.63
Garg, N., Gupta, A., Kumar, A., Singla, S.: Non-clairvoyant precedence constrained scheduling. In: Baier, C., Chatzigiannakis, I., Flocchini, P., Leonardi, S. (eds.) 46th International Colloquium on Automata, Languages and Programming (ICALP). LIPIcs, vol. 132, pp. 63:1–63:14. Dagstuhl (2019). https://doi.org/10.4230/LIPIcs.ICALP.2019.63
Lassota, A.A., Lindermayr, A., Megow, N., Schlöter, J.: Minimalistic predictions to schedule jobs with online precedence constraints. In: Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., Scarlett, J. (eds.) Proceeding of 40th International Conference on Machine Learning, PMLR, vol. 202, pp. 18563–18583. https://proceedings.mlr.press/v202/lassota23a.html
Lindermayr, A., Megow, N.: Permutation predictions for non-clairvoyant scheduling. In: Lee, I.-T.A. (ed.) Proceeding of 34th Symposium on Parallel Algorithms Architecture (SPAA), pp. 357–368. ACM, New York, NY (2022). https://doi.org/10.1145/3490148.3538579
Megow, N., Vredeveld, T.: A tight 2-approximation for preemptive stochastic scheduling. Math. Oper. Res. 39(4), 1297–1310 (2014). https://doi.org/10.1287/moor.2014.0653
Sevcik, K.C.: Scheduling for minimum total loss using service time distributions. J. ACM 21(1), 66–75 (1974). https://doi.org/10.1145/321796.321803
Weiss, G.: On almost optimal priority rules for preemptive scheduling of stochastic jobs on parallel machines. Adv. Appl. Probab. 27(3), 821–839 (1995). https://doi.org/10.2307/1428135
Graham, R.L.: Bounds for certain multiprocessing anomalies. Bell Syst. Tech. J. 45(9), 1563–1581 (1966). https://doi.org/10.1002/j.1538-7305.1966.tb01709.x
Conway, R.W., Maxwell, W.L., Miller, L.W.: Theory of Scheduling. Addison-Wesley, Reading, Palo Alto, London, Don Mills (1967)
Rinnooy Kan, A.H.G.: Machine Scheduling Problems: Classification, Complexity and Computations. Martinus Nijhoff, Den Haag (1976). https://doi.org/10.1007/978-1-4613-4383-7
Afrati, F., Bampis, E., Chekuri, C., Karger, D., Kenyon, C., Khanna, S., Milis, I., Queyranne, M., Skutella, M., Stein, C., Sviridenko, M.: Approximation schemes for minimizing average weighted completion time with release dates. In: 40th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pp. 32–43 (1999). https://doi.org/10.1109/SFFCS.1999.814574
Hoogeveen, J.A., Vestjens, A.P.A.: Optimal on-line algorithms for single-machine scheduling. In: Cunningham, W.H., McCormick, S.T., Queyranne, M. (eds.) Integer Programming Combined Optimization (IPCO). LNCS, vol. 1084, pp. 404–414. Springer, Berlin, Heidelberg (1996). https://doi.org/10.1007/3-540-61310-2_30
Anderson, E.J., Potts, C.N.: Online scheduling of a single machine to minimize total weighted completion time. Math. Oper. Res. 29(3), 686–697 (2004). https://doi.org/10.1287/moor.1040.0092
Möhring, R.H., Radermacher, F.J., Weiss, G.: Stochastic scheduling problems I - general strategies. Z. Oper. Res. 28(7), 193–260 (1984). https://doi.org/10.1007/BF01919323
Scharbrodt, M., Schickinger, T., Steger, A.: A new average case analysis for completion time scheduling. J. ACM 53(1), 121–146 (2006). https://doi.org/10.1145/1120582.1120585
Rothkopf, M.H.: Scheduling with random service times. Manag. Sci. 12(9), 707–713 (1966). https://doi.org/10.1287/mnsc.12.9.707
Möhring, R.H., Schulz, A.S., Uetz, M.: Approximation in stochastic scheduling: the power of LP-based priority policies. J. ACM 46(6), 924–942 (1999). https://doi.org/10.1145/331524.331530
Jäger, S.J.: Approximation in deterministic and stochastic machine scheduling. PhD thesis, Technische Universität Berlin (2021). https://doi.org/10.14279/depositonce-12198
Jäger, S., Skutella, M.: Generalizing the Kawaguchi-Kyan bound to stochastic parallel machine scheduling. In: Niedermeier, R., Vallée, B. (eds.) 35th International Symposium on Theoretical Aspects of Computer Science (STACS). LIPIcs, vol. 96, pp. 43:1–43:13. Dagstuhl (2018). https://doi.org/10.4230/LIPIcs.STACS.2018.43
Im, S., Moseley, B., Pruhs, K.: Stochastic scheduling of heavy-tailed jobs. In: Mayr, E.W., Ollinger, N. (eds.) 32nd International Symposium on Theoretical Aspects of Computer Science (STACS). LIPIcs, vol. 30, pp. 474–486. Dagstuhl (2015). https://doi.org/10.4230/LIPIcs.STACS.2015.474
McNaughton, R.: Scheduling with deadlines and loss functions. Manag. Sci. 6(1), 1–12 (1959). https://doi.org/10.1287/mnsc.6.1.1
Epstein, L., Levin, A.: The benefit of preemption for single machine scheduling so as to minimize total weighted completion time. Oper. Res. Lett. 44(6), 772–774 (2016). https://doi.org/10.1016/j.orl.2016.09.013
Labetoulle, J., Lawler, E.L., Lenstra, J.K., Rinnooy Kan, A.H.G.: Preemptive scheduling of uniform machines subject to release dates. In: Pulleyblank, W.R. (ed.) Progress in Combinatorial Optimization, pp. 245–261. Academic Press, Don Mills, ON (1984). https://doi.org/10.1016/B978-0-12-566780-7.50020-9
Sitters, R.: Competitive analysis of preemptive single-machine scheduling. Oper. Res. Lett. 38(6), 585–588 (2010). https://doi.org/10.1016/j.orl.2010.08.012
Epstein, L., van Stee, R.: Lower bounds for on-line single-machine scheduling. Theor. Comput. Sci. 299(1), 439–450 (2003). https://doi.org/10.1016/S0304-3975(02)00488-7
Parekh, A.K.J.: A generalized processor sharing approach to flow control in integrated services networks. PhD thesis, Massachusetts Institute of Technology (1992). https://dspace.mit.edu/handle/1721.1/13076
Chandra, A., Adler, M., Goyal, P., Shenoy, P.: Surplus fair scheduling: A proportional-share CPU scheduling algorithm for symmetric multiprocessors. In: 4th Symposium on Operating Systems Design and Implementation (OSDI), pp. 4:1–4:14. USENIX, Berkeley, CA (2000). https://www.usenix.org/conference/osdi-2000/surplus-fair-scheduling-proportional-share-cpu-scheduling-algorithm-symmetric
Becchetti, L., Leonardi, S.: Nonclairvoyant scheduling to minimize the total flow time on single and parallel machines. J. ACM 51(4), 517–539 (2004). https://doi.org/10.1145/1008731.1008732
Kalyanasundaram, B., Pruhs, K.: Speed is as powerful as clairvoyance. J. ACM 47(4), 617–643 (2000). https://doi.org/10.1145/347476.347479
Bansal, N., Dhamdhere, K.: Minimizing weighted flow time. ACM Trans. Algor. 3(4), 39–13914 (2007). https://doi.org/10.1145/1290672.1290676
Bansal, N., Pruhs, K.: Server scheduling in the weighted \(\ell _p\)-norm. In: Farach-Colton, M. (ed.) Theor. Inform. (LATIN). LNCS, vol. 2976, pp. 434–443. Springer, Berlin, Heidelberg (2004). https://doi.org/10.1007/978-3-540-24698-5_47
Vestjens, A.P.A.: On-line machine scheduling. PhD thesis, Technische Universiteit Eindhoven (November 1997).https://doi.org/10.6100/IR500043
Stee, R., La Poutré, H.: Minimizing the total completion time on-line on a single machine, using restarts. J. Algor. 57(2), 95–129 (2005). https://doi.org/10.1016/j.jalgor.2004.10.001
Chekuri, C., Motwani, R., Natarajan, B., Stein, C.: Approximation techniques for average completion time scheduling. SIAM J. Comput. 31(1), 146–166 (2001). https://doi.org/10.1137/S0097539797327180
Dürr, C., Erlebach, T., Megow, N., Meißner, J.: An adversarial model for scheduling with testing. Algorithmica 82(12), 3630–3675 (2020). https://doi.org/10.1007/s00453-020-00742-2
Albers, S., Eckl, A.: Explorable uncertainty in scheduling with non-uniform testing times. In: Kaklamanis, C., Levin, A. (eds.) Approximation and Online Algorithms (WAOA). LNCS, vol. 12806, pp. 127–142. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-80879-2_9
Gong, M., Lin, G.: Improved approximation algorithms for multiprocessor scheduling with testing. In: Chen, J., Li, M., Zhang, G. (eds.) Frontiers Algorithmics (IJTCS-FAW). LNCS, vol. 12874, pp. 65–77. Springer, Cham (2021).https://doi.org/10.1007/978-3-030-97099-4_5
Baeza-Yates, R.A., Culberson, J.C., Rawlins, G.J.E.: Searching in the plane. Inf. Comput. 106(2), 234–252 (1993). https://doi.org/10.1006/inco.1993.1054
Beck, A., Newman, D.J.: Yet more on the linear search problem. Israel J. Math. 8(4), 419–429 (1970). https://doi.org/10.1007/BF02798690
Kao, M.-Y., Reif, J.H., Tate, S.R.: Searching in an unknown environment: an optimal randomized algorithm for the cow-path problem. Inf. Comput. 131(1), 63–79 (1996). https://doi.org/10.1006/inco.1996.0092
Kao, M.-Y., Ma, Y., Sipser, M., Yin, Y.: Optimal constructions of hybrid algorithms. J. Alg. 29(1), 142–164 (1998). https://doi.org/10.1006/jagm.1998.0959
Yao, A.C.-C.: Probabilistic computations: Toward a unified measure of complexity. In: 18th Annual IEEE Symposium on Foundations of Computer Science (SFCS), pp. 222–227 (1977). https://doi.org/10.1109/SFCS.1977.24
Böttcher, A., Grudsky, S.M.: Spectral Properties of Banded Toeplitz Matrices. SIAM, Philadelphia, PA (2005)
Goemans, M.X.: A supermodular relaxation for scheduling with release dates. In: Cunningham, W.H., McCormick, S.T., Queyranne, M. (eds.) Integer Programming and Combinatorial Optimization. LNCS, vol. 1084, pp. 288–300. Springer, Berlin, Heidelberg (1996). https://doi.org/10.1007/3-540-61310-2_22
Goemans, M.X.: Improved approximation algorthims for scheduling with release dates. In: Saks, M. (ed.) Proceeding of Eighth Annual of ACM-SIAM Symposium on Discrete Algorithms (SODA), pp. 591–598 (1997).https://doi.org/10.5555/314161.314394
Acknowledgements
We thank Sungjin Im for helpful comments on an earlier version of this manuscript. The research of the second, third and fourth authors was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy—The Berlin Mathematics Research Center MATH+ (EXC-2046/1, project ID: 390685689).
Funding
Open Access funding enabled and organized by Projekt DEAL. The research of the second, third and fourth authors was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy—The Berlin Mathematics Research Center MATH+ (EXC-2046/1, project ID: 390685689).
Author information
Authors and Affiliations
Contributions
All authors contributed in equal parts.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no relevant financial or non-financial interests to disclose.
Ethical approval
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
An extended abstract of this paper appeared in the proceedings of IPCO 2023 [1].
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Jäger, S., Sagnol, G., Schmidt genannt Waldschmidt, D. et al. Competitive kill-and-restart and preemptive strategies for non-clairvoyant scheduling. Math. Program. 210, 457–509 (2025). https://doi.org/10.1007/s10107-024-02118-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10107-024-02118-8