1 Introduction

Solving stiff initial value problems (IVPs) efficiently and accurately remains a significant challenge in numerical analysis. Explicit Runge–Kutta (RK) method, face limitations when applied to stiff systems due to their stability restrictions. To address these limitations, a new class of time-advancing schemes for the numerical solution of stiff IVPs was proposed by Bassenne, Fu, and Mani in [1]. They introduced the concept of TASE (Time-Accurate and Stable Explicit) operators, designed to enhance the stability of explicit RK methods. In this approach, instead of solving the differential system

$$\begin{aligned} \dfrac{d}{dt} Y(t) = F ( t, Y(t)), \quad Y(t_0)=Y_0 \in \mathbb {R}^d. \end{aligned}$$
(1)

they proposed to solve another (stabilized) IVP

$$\begin{aligned} \dfrac{d}{dt} U(t) = T \; F ( t, U(t)), \quad U(t_0)=U_0 \equiv Y_0 \in \mathbb {R}^d, \end{aligned}$$
(2)

where \( T = T( \varDelta t)\in \mathbb {R}^{d\times d}\) is a linear operator that may depend on the time step size \(\varDelta t >0\) and on F, so that if \(U_{RK}(t_n)\) is the numerical approximation at \(t_n\) of the solution of Eq. (2) obtained by integrating it with an explicit RK method of order p, then it satisfies \( U_{RK}(t_0 + \varDelta t) - Y(t_0 + \varDelta t) = \mathcal {O}(\varDelta t ^{p+1}), \) i.e., approximates the local solution of Eq. (1) with order p, and also satisfies some stability requirements that are necessary for solving stiff systems such as A- or L- stability. In this way, the introduction of the TASE operator into the original governing Eq. (1) allows us to overcome the numerical stability restrictions of explicit RK time-advancing methods for solving stiff systems.

In terms of the accuracy order of the numerical solution of Eq. (2), if the explicit RK scheme has order p and if the TASE operator satisfies

$$\begin{aligned} T ( \varDelta t ) = I + \mathcal {O} ( \varDelta t^q), \end{aligned}$$
(3)

it can be easily proved that the numerical solution of the stabilized system Eq. (2), \(U_{RK}(t_n)\), provided by an explicit RK method of order p has at least order \(\min (p,q)\). Some specific schemes combining explicit p–stage RK methods with orders \( p \le 4\) and TASE operators with the same order were derived in [1].

A more general family of TASE operators was derived in [2] by taking

$$\begin{aligned} T(\varDelta t)= \sum _{j=1}^p \beta _j (I-\varDelta t\; \alpha _j W)^{-1}, \end{aligned}$$
(4)

where W is an approximation to the Jacobian matrix of the vector field F at the current integration step \((t_n, U_{RK}(t_n))\), \(\alpha _j >0\), \( j=1, \ldots ,p\), are free real parameters, and \(\beta _j\) are uniquely determined by the condition \( T_p ( \varDelta t ) = I + \mathcal {O} ( \varDelta t^p) \). The free parameters \(\alpha _j\) were selected to improve the linear stability properties of an explicit RK method with order p for the scaled equation Eq. (2). Further generalization of TASE operators were given in [3], where complex coefficients were allowed in the operator, yet working in real arithmetic.

In the TASE operators described by Eq. (4), each evaluation of T involves the solution of p linear systems with matrices \( ( I - \varDelta t\;\alpha _j W)\), \( j=1, \ldots ,p\). If these systems are to be solved by some LU matrix factorization, it would be much more efficient if the \(\alpha _j\) coefficients were all equal, but this is not compatible with order p for the TASE operator.

To make the methods more efficient, Calvo et al. [4] developed a different family of TASE operators, termed Singly-TASE operators, defined by

$$\begin{aligned} T_p(\varDelta t)= \sum _{j=1}^p \beta _j (I-\varDelta t\; \alpha _j W)^{-j}. \end{aligned}$$
(5)

These operators lead to significant improvements in computational efficiency while maintaining the stability and accuracy necessary for solving stiff problems.

For an s–stage explicit RK method defined by the Butcher tableau

$$\begin{aligned} \begin{array}{c|ccccc} 0 & & & & & \\ c_2 & a_{21} & & & & \\ \vdots & & & & & \\ c_s & a_{s1} & a_{s2} & \ldots & a_{s,s-1} & \\ \hline & b_1 & b_2 & \ldots & b_{s-1} & b_s \end{array} \qquad c_i = \sum _{j=1}^{i-1} a_{ij}, \quad i= 2, \ldots , s, \end{aligned}$$

the numerical solution of Eq. (2) by a Singly-RKTASE method is advanced from \( (t_0, U_0) \rightarrow (t_1= t_0 + \varDelta t, U_1)\) by the formula

$$\begin{aligned} U_1 = U_0 + \varDelta t \left[ b_1 \; K_1 + \ldots + b_s \; K_s \right] , \end{aligned}$$

where \( K_j \), \( j=1, \ldots ,s\) are computed recursively from the formulas

$$\begin{aligned} \begin{array}{lll} K_1 & =& T \; F(t_0, U_0), \\ K_2 & =& T \; F(t_0 + c_2 \varDelta t, U_0 + \varDelta t \; a_{21} K_1), \\ & \vdots & \\ K_s & =& T \; F \left( t_0 + c_s \varDelta t, U_0 + \varDelta t \; \sum _{j=1}^{s-1} a_{s j} K_j \right) . \end{array} \end{aligned}$$
(6)

In this paper, we propose a further modification of the Singly-RKTASE methods to enhance efficiency, so that for the ith stage of the RK method we use a TASE operator defined by

$$\begin{aligned} T_i = \sum _{j=1}^r \beta _{ij} \; ( I - \varDelta t \; \alpha W)^{-j}, \quad \alpha >0. \end{aligned}$$
(7)

The coefficients \(\alpha >0\), r and \( \beta _{ij}\), \( j=1, \ldots ,r\) must be determined so that the resulting modified singly-RKTASE method

$$\begin{aligned} \begin{array}{lll} K_1 & =& T_1 \; F(t_0, U_0), \\ K_2 & =& T_2 \; F(t_0 + c_2 \varDelta t, U_0 + \varDelta t \; a_{21} K_1), \\ & \vdots & \\ K_s & =& T_s \; F \left( t_0 + c_s \varDelta t, U_0 + \varDelta t \; \sum _{j=1}^{s-1} a_{s j} K_j \right) , \\ U_1 & =& U_0 + \varDelta t [ b_1 K_1 + \ldots + b_s K_s] \end{array} \end{aligned}$$
(8)

has order p. Compared to Singly-RKTASE methods, instead of s coefficients \(\beta _i\) we have sr coefficients \(\beta _{ij}\), providing more freedom to improve accuracy and stability properties. Although we could take different \(r_i\) for each stage, we have taken the same r for simplicity. In order to get order p for the resulting MSRKTASE method, we can impose, as with Singly-RKTASE methods, that all operators \(T_i\) in (7) have order p. However, this is only achieved if \(r \ge p\). For \(r=p\) the coefficients \(\beta _{ij}\) are uniquely determined, resulting in a Singly-RKTASE and if \(r>p\) we get more freedom in the coefficients, but at the price of increasing the computational cost. Then, we will consider operators \(T_i\) with order \(q < p\). In this case, only order \(\min (q,p)<p\) is guaranteed. Fortunately, we will see that if \(r=p\) it is possible to get final order p with operators \(T_i\) with orders \(q<p\). This will be done by studying the conditions that the methods’s coefficients must satisfy to achieve order p, using the order theory of W-methods [5], following the ideas in [6].

The rest of the paper is organized as follows. In Section 2, we prove that modified Singly-RKTASE methods are a particular case of W-methods. In Section 3, we derive the conditions that the coefficients \(\beta _{ij}\) and \(\alpha \) of the MSRKTASE must satisfy to have order p. In section 4 we develop MSRKTASE methods of orders 2 and 3 with optimal stability and accuracy properties. Finally, in section 5, through numerical experiments, we show the effectiveness of the newly developed methods, highlighting their potential for solving both linear and nonlinear stiff systems with improved performance.

2 Connection Between Modified Singly-RKTASE and W-Methods

Given a matrix W (usually some approximation to the Jacobian matrix of the vector field F(tY) at the current integration point \((t_n, Y_n)\)) a singly W-method [5] with \(\hat{s}\) stages provides an approximation to the solution of Eq. (1) at \(t_{n+1}=t_n + h\) by the formulas

$$\begin{aligned} \begin{array}{l} (I-h\alpha W) {\widehat{K}}_1 =h f(Y_n), \\ (I-h\alpha W) {\widehat{K}}_i = h f( Y_n+ \sum _{j=1}^{i-1} {\hat{a}}_{ij} {\widehat{K}}_j)+ h\alpha W \sum _{j=1}^{i-1} l_{ij} {\widehat{K}}_j, \quad i=2, \ldots , {\hat{s}} \\ Y_{n+1} = Y_n + \sum _{i=1}^{{\hat{s}}} {\hat{b}}_i {\widehat{K}}_i. \end{array} \end{aligned}$$

The coefficients of the method can be arranged in a Butcher tableau as

On the other side, the product of the operator \(T_i\) in (7) by a vector v can be written as

$$\begin{aligned} \begin{array}{l} w_1 = (I-h\alpha W)^{-1} v, \\ w_j =(I-h\alpha W)^{-1} w_{j-1}, \quad j=2, \ldots , r \\ T_i v = \beta _{i1} w_1 + \ldots + \beta _{ir} w_r. \end{array} \end{aligned}$$

Then, the Eqs. (7)(8) defining a Modified Singly-RKTASE scheme can be written as

$$\begin{aligned} \begin{array}{l} (I-h\alpha W) {\widehat{K}}_1 = h f(Y_n), \\ (I-h\alpha W) {\widehat{K}}_j = {\widehat{K}}_{j-1}, \quad j=2, \ldots , r, \\ K_1= \beta _{11} {\widehat{K}}_1 + \ldots + \beta _{1r} {\widehat{K}}_r, \\ (I-h\alpha W) {\widehat{K}}_{r(i-1)+1} = h f( Y_n+ \sum _{j=1}^{i-1} a_{ij} K_j), \qquad i= 2, \ldots ,s,\\ (I-h\alpha W) {\widehat{K}}_{r(i-1)+j} = {\widehat{K}}_{r(i-1)+j-1}, \quad j=2, \ldots , r,\\ K_i= \beta _{i1} {\widehat{K}}_{r(i-1)+1} + \ldots + \beta _{ir} {\widehat{K}}_{r i }, \\ Y_{n+1}= Y_n + \sum _{i=1}^s b_i K_i. \end{array} \end{aligned}$$

Taking into account that \( {\widehat{K}}_j = {\widehat{K}}_{j-1} + h \alpha W {\widehat{K}}_j, \) we have

$$\begin{aligned} \begin{array}{l} (I-h\alpha W) {\widehat{K}}_1 = h f(Y_n), \\ (I-h\alpha W) {\widehat{K}}_j = h f(Y_n) + h\alpha W \sum _{l=1}^{j-1} {\widehat{K}}_{l}, \quad j=2, \ldots , r, \\ K_1= \beta _{11} {\widehat{K}}_1 + \ldots + \beta _{1r} {\widehat{K}}_r, \\ (I-h\alpha W) {\widehat{K}}_{r(i-1)+1} = h f( Y_n+ \sum _{j=1}^{i-1} a_{ij} K_j), \quad i= 2, \ldots ,s,\\ (I-h\alpha W) {\widehat{K}}_{r(i-1)+j} = h f( Y_n+ \sum _{j=1}^{i-1} a_{ij} K_j) + \\ \hspace{102.43008pt}h\alpha W \sum _{l=1}^{j-1} {\widehat{K}}_{r(i-1)+l-1}, \quad j=2, \ldots , r, \\ K_i= \beta _{i1} {\widehat{K}}_{r(i-1)+1} + \ldots + \beta _{ir} {\widehat{K}}_{r i }, \\ Y_{n+1}= Y_n + \sum _{i=1}^s b_i K_i. \end{array} \end{aligned}$$

and this leads to

$$\begin{aligned} \begin{array}{l} (I-h\alpha W) {\widehat{K}}_1 = h f(Y_n), \\ (I-h\alpha W) {\widehat{K}}_j = h f(Y_n) + h\alpha W \sum _{l=1}^{j-1} {\widehat{K}}_{l}, \quad j=2, \ldots , r, \\ (I-h\alpha W) {\widehat{K}}_{r(i-1)+j} = h f( Y_n+ \sum _{j=1}^{i-1} a_{ij} \sum _{l=1}^r \beta _{jl} {\widehat{K}}_{r(i-1)+j}) + \\ \hspace{102.43008pt}h\alpha W \sum _{l=1}^{j-1} {\widehat{K}}_{r(i-1)+l-1}, \quad i=2, \ldots , s, \quad j=2, \ldots , r, \\ Y_{n+1}= Y_n + \sum _{i=1}^s b_i \sum _{l=1}^r \beta _{il} \widehat{K}_{r(i-1)+j}. \end{array} \end{aligned}$$

This is just an W-method with \({\hat{s}}=sr\) stages whose vector \({\widehat{b}}\) and matrices \({\widehat{A}}\), L are given by

$$\begin{aligned} L= \begin{pmatrix} L_r & & \\ & \ddots \\ & & L_r \end{pmatrix}\in \mathbb {R}^{rs\times rs}, \qquad {\hat{A}} = \begin{pmatrix} 0 & & \\ A_{21} & 0\\ \vdots & \\ A_{s1} & \cdots & A_{s,s-1} & 0 \end{pmatrix}\in \mathbb {R}^{rs\times rs}, \end{aligned}$$
$$\begin{aligned} {\widehat{b}} = (b_1 \beta _{11},\ldots , b_1 \beta _{1r}, \ldots , b_s \beta _{s1},\ldots , b_s \beta _{sr})^T, \end{aligned}$$

with

$$\begin{aligned} L_r= \begin{pmatrix} 0 & & \\ 1& 0 & \\ \vdots & \ddots & \ddots \\ 1 & \cdots & 1 & 0 \end{pmatrix}\in \mathbb {R}^{r\times r}, \qquad A_{ij} = a_{ij} \begin{pmatrix} \beta _{j1} & \ldots & \beta _{jr} \\ \vdots & \\ \beta _{j1} & \ldots & \beta _{jr} \end{pmatrix}\in \mathbb {R}^{r\times r}. \end{aligned}$$

This equivalence allows us to analyze the order of Modified Singly-RKTASE methods through the order conditions of W-methods. Also, the absolute stability can be studied through the W-methods. The stability function of a W-method is given by [5, 7]

$$\begin{aligned} {\hat{R}}(z)= 1 + z {\hat{b}}^T (I-z({\hat{A}} + \varGamma ))^{-1} \textbf{1} \end{aligned}$$

and the limit of this function when z goes to infinity is given by

$$\begin{aligned} \lim _{z\rightarrow \infty } {\hat{R}}(z) = 1-{\hat{b}}^T({\hat{A}} + \varGamma )^{-1} \textbf{1}. \end{aligned}$$
(9)

We will use this to develop Modified Singly-RKTASE methods. Recall that

  • A method is called A-stable if \(\vert {\hat{R}}( z)\vert \le 1\) for all Re \(z\le 0\);

  • A method is said to be L-stable if it is A-stable and \(\hat{R}(\infty )=0\);

  • A method is called \(A(\theta )\)-stable if \(\vert {\hat{R}}(z)\vert <1\) for all z such that \(\arg (-z) \le \theta \), that is, its stability region contains the left hand region of the complex plane with angle \(\theta \). If in addition \({\hat{R}}(\infty )=0\), it is called \(L(\theta )\)-stable.

3 Order Conditions

Denoting \(\varGamma = \alpha (I+L)\in \mathbb {R}^{{\hat{s}}\times {\hat{s}}}\), and \(\textbf{1}=(1,\ldots , 1)^T\in \mathbb {R}^{{\hat{s}}}\), the order conditions for a W-method are given in Table 1 (see e.g. [5, 7]).

Table 1 Order conditions for W-methods up to order 4

In the case the operator W is exactly the Jacobian matrix, \(W=\partial f/\partial y (t_n, Y_n)\) (Rosenbrock-Wanner methods [8]), many elementary differentials become the same and the order conditions reduce to those in Table 2.

Table 2 Order conditions for a Rosenbrock method (\(W=\partial f/\partial y\)) up to order 4

Let us analyze the order conditions for a Modified Singly-RKTASE scheme expressed as its equivalent W-method. We can take without losing generality

$$\begin{aligned} \beta _{i1}+ \ldots + \beta _{ir}=1, \quad i=1, \ldots , s \end{aligned}$$
(10)

because this is equivalent to re-scale the coefficients bA of the underlying RK method.

First, the order conditions that do not involve the matrix \(\varGamma \) reduce to the order conditions associated to the undelying RK method. For example, \({\hat{b}}^T \textbf{1}=b_1+ \ldots +b_s\), or \(\hat{b}^T {\hat{A}} {\hat{c}} = b^T A c\) which lead to the corresponding equations of the RK method.

On the other hand, the form of the matrix L introduces some incompatibility between some of the order conditions leading to the next theorem.

Theorem 1

A Modified Singly-RKTASE method can not have order \(r+1\).

Proof

Since the matrix \(\varGamma \) is a block diagonal matrix with blocks \(\alpha (I_r+L_r)\), it is clear that \((\varGamma - \alpha I)^r =(\alpha L)^r=0\) because \(L_r\) is strictly lower triangular with dimension r. Then, \((\varGamma - \alpha I)^r \textbf{1}=0\). If the method had order \(r+1\), by the order conditions it should be \(\widehat{b}^T(\varGamma - \alpha I)^r \textbf{1}=\alpha ^r\) which is not zero if \(\alpha \ne 0\). \(\square \)

With this result, a Modified Singly-RKTASE method with order p must have necessarily \(r \ge p\).

4 Modified Singly-RKTASE Methods with Orders 2 and 3

4.1 Methods with \(s=2\), \(r=2\) and order 2

There exists a family of Runge–Kutta schemes with two stages and order two that depend on one coefficient \(c_2\) given by

$$\begin{aligned} a_{21}=c_2,\quad b_2=\dfrac{1}{2 c_2}, \quad b_1=1-b_2. \end{aligned}$$

Since \(b^T A c=0\) for any \(c_2\), the associated order condition of order three does not depend on \(c_2\) and the other one vanishes for \(c_2= 2/3\). So, we will take this value of \(c_2\).

The only remaining condition for a Modified Singly-RKTASE with \(s=r=2\) to achieve order 2 is \({\hat{b}}^T \varGamma \textbf{1} =0\) which is satisfied for

$$\begin{aligned} \beta _{22}= -(4+\beta _{12})/3, \end{aligned}$$

and \(\beta _{11}=1-\beta _{12}, \beta _{21}=1-\beta _{22}\) according to (10). The limit of the stability function when z approaches infinity for the case of our Modified Singly-RKTASE method of order two is given by

$$\begin{aligned} {\hat{R}}(\infty )= \lim _{z\rightarrow \infty } {\hat{R}}(z) = 1-{\hat{b}}^T({\hat{A}} + \varGamma )^{-1} \textbf{1} = -\dfrac{-7 + 12 \alpha - 6 \alpha ^2 + 6 \beta _{12} + \beta _{12}^2}{6 \alpha ^2}. \end{aligned}$$

There are two values of \(\beta _{12}\) that make \({\hat{R}}(\infty )=0\). The one that yields better results in terms of stability and accuracy is

$$\begin{aligned} \beta _{12}= -3 + \sqrt{16 - 12 \alpha + 6 \alpha ^2}. \end{aligned}$$

There is only one free parameter \(\alpha \) left, which we will use to maximize the stability region and minimize the error coefficients. The 2-norm of the error coefficients of order 3 is given by

$$\begin{aligned} C_3=\left[ ({\hat{b}}^T{\hat{A}} {\hat{c}}-1/6)^2+ ({\hat{b}}^T \varGamma ^2 \textbf{1})^2+ ({\hat{b}}^T {\hat{A}} \varGamma \textbf{1})^2+ ({\hat{b}}^T \varGamma {\hat{A}} \textbf{1})^2 \right] ^{1/2}. \end{aligned}$$

The corresponding norm when \(W= F'\) is

$$\begin{aligned} D_3= \left| {\hat{b}}^T(\varGamma +{\hat{A}})^2 \textbf{1}-1/6 \right| . \end{aligned}$$

The method is A-stable (and therefore L-stable) if \(\alpha \in [ 0.3117, 3.257]\) and \(C_3\) and \(D_3\) increase monotonically when \(\alpha \) increases, so the smallest value of \(\alpha \) is more convenient. A good value of the parameter is \(\alpha =32/100\). With this value, we obtain an L-stable method of order 2, which we will name MSRKTASE2. The error coefficients are given in Table 3. For comparison, we also include in this Table the error coefficients of the Singly-RKTASE of order 2 obtained in [4], named SRKTASE2. As we can see, the new method achieves L-stability (SRKTASE2 has \({\hat{R}}(\infty )=1/2\)) and error coefficients that are about 20 times smaller (40 times smaller in the case where \(W=f'\)).

We could have used the parameter \(\beta _{12}\) to obtain smaller error coefficients without imposing the condition \(R(\infty )=0\). For example we could nullify the coefficient \(D_3\) (order three for the case \(W=f'\)) but in this case, we only achieve reasonable stability regions for large values of \(\alpha \) and \(C_3\) becomes very large (greater than one hundred).

The coefficients of the obtained method, MSRKTASE2, are given in the Appendix.

Table 3 Properties of the proposed methods

4.2 Methods with \(s=3\), \(r=3\) and Order 3

There exist a family of Runge–Kutta schemes with three stages and order three, defined by the coefficients

$$\begin{aligned} a_{32}=\dfrac{c_3(c_2-c_3)}{c_2(3c_2-2)},\quad b_2=\dfrac{2 - 3 c_3}{6 c_2 (c_2 - c_3)}, \quad b_3=\dfrac{2 - 3 c_2}{6 c_3 (c_3 - c_2)}, \quad b_1=1-b_2-b_3, \end{aligned}$$

depending on two free parameters \(c_2\) and \(c_3\) with \(c_2\ne c_3\) and \(c_2 \ne 2/3\).

The remaining conditions for a Modified Singly-RKTASE with \(s=r=3\) to achieve order 3 form the set of four equations \({\hat{b}}^T \varGamma \textbf{1} =0\), \({\hat{b}}^T \varGamma ^2 \textbf{1} =0\), \({\hat{b}}^T {\hat{A}} \varGamma \textbf{1} =0 \), \({\hat{b}}^T \varGamma {\hat{A}} \textbf{1} =0\) which can be solved for the parameters \(\beta _{12}, \beta _{13}, \beta _{23}\) and \(\beta _{33}\) (\(\beta _{11}, \beta _{21}\) and \(\beta _{31}\) are obtained from (10)), resulting in:

$$\begin{aligned} \begin{array}{l} \beta _{12}=\dfrac{c_3 (-2 + 3 c_3) \beta _{22} - 3 c_2^2 (6 c_3 + \beta _{32}) + 2 c_2 (9 c_3^2 + \beta _{32})}{(c_2 - c_3) (2 - 3 c_3 + c_2 (-3 + 6 c_3))}, \\ \beta _{13}= -\dfrac{c_3 (-2 + 3 c_3) (1 + \beta _{22}) - 3 c_2^2 (1 + 4 c_3 + \beta _{32}) + 2 c_2 (1 + 6 c_3^2 + \beta _{32})}{2 (c_2 - c_3) (2 - 3 c_3 + c_2 (-3 + 6 c_3))}\\ \beta _{23} = \dfrac{1}{2} (-1 - \beta _{22}), \\ \beta _{33} = \dfrac{1}{2} (-1 - \beta _{32}), \end{array} \end{aligned}$$

depending on the parameters \(c_2, c_3, \beta _{22}\) and \(\beta _{32}\). Note that in this family of methods of order three, \(\alpha \) is also a free parameter. Imposing \(\beta _{32}=\beta _{22}=\beta _{12}\) we obtain the third order Singly-RKTASE methods obtained in [4].

The free parameters can be selected to maximize the stability region and to minimize the coefficients of the leading term of the local error. For these methods of order three, it is also satisfied that

$$\begin{aligned} \begin{array}{l} {\hat{b}}^T{\hat{A}} \varGamma {\hat{A}} \textbf{1}=0, \quad {\hat{b}}^T{\hat{A}}^2 \varGamma \textbf{1}=0, \quad {\hat{b}}^T \varGamma {\hat{A}}^2 \textbf{1}=0,\\ {\hat{b}}^T \varGamma {\hat{A}} \varGamma \textbf{1}=0, \quad {\hat{b}}^T \varGamma {\hat{c}}^2=0,\quad {\hat{b}}^T({\hat{A}}\varGamma \textbf{1}) \cdot {\hat{c}}=0, \end{array} \end{aligned}$$

and the other error coefficients of the term of order four are

$$\begin{aligned} & C_{41}\equiv {\hat{b}}^T{\hat{A}}^2 {\hat{c}} -1/24 = b^T A^2 c -1/24 = -1/24, \\ & C_{42}\equiv (8{\hat{b}}^T ({\hat{A}} {\hat{c}} \cdot {\hat{c}}) -1)/24 = (8 b^T (A c \cdot c)-1)/24 =(4c_3-3)/72, \\ & C_{43}\equiv (12 {\hat{b}}^T{\hat{A}} {\hat{c}}^2 -1)/24 = (12 b^T A c^2 -1)/24 =(2c_2-1)/24, \\ & C_{44}\equiv (4{\hat{b}}^T{\hat{c}}^3-1)/24 = (4b^T c^3-1)/24 = (2((c_2 (2 - 3 c_3) + 2 c_3)-3)/72, \end{aligned}$$
$$\begin{aligned} & C_{45}\equiv {\hat{b}}^T{\hat{A}} \varGamma ^2 \textbf{1}\nonumber \\ & \quad =\alpha ^2 \frac{-2 \beta _{22} \!+ \!6 c_3 (3 \!+ \!\beta _{22}) \!-\! 3 c_3^2 (3 \!+ \!\beta _{22}) \!+\! 2 \beta _{32} \!+\! 9 c_2^2 (3 \!+ \!\beta _{32}) \!-\! 3 c_2 (6 \!- \!\beta _{22} \!+ \!2 c_3 (3 \!+ \!\beta _{22}) \!+ \!3 \beta _{32}))}{12 (c_2 \!-\! c_3) (2 \!-\! 3 c_3 \!+\! c_2 (-3 \!+\! 6 c_3))}, \nonumber \\ & C_{46}\equiv {\hat{b}}^T \varGamma ^2 {\hat{c}}= \alpha ^2 \dfrac{(-2 \beta _{22} + 3 c_3 (3 + \beta _{22}) + 2 \beta _{32} - 3 c_2 (3 + \beta _{32}))}{12 (c_2 - c_3)}, \nonumber \\ & C_{47}\equiv {\hat{b}}^T \varGamma ^3 \textbf{1}= \alpha ^3. \end{aligned}$$
(11)

If we solve \({\hat{b}}^T \varGamma ^2 {\hat{c}}=0\) and \({\hat{b}}^T{\hat{A}} \varGamma ^2 \textbf{1}=0\) for \(\beta _{22}\) and \(\beta _{32}\) we obtain the method in [4], independent of the values of \(c_2\) and \(c_3\).

The first four equations in (11) depend only on \(c_2\) and \(c_3\) and can not be satisfied simultaneously. The 2-norm of these four equations reaches its minimum value at \(c_2= 0.496188\), \(c_3=0.764887\), very close to \(c_2=1/2\), \(c_3=3/4\) for which

$$\begin{aligned} \begin{array}{l} \dfrac{\big ((24 b^T A^2c-1)^2+(12b^T Ac^2-1)^2+(8b^T (Ac\cdot c)-1)^2+(4b^T c^3-1)^2\big )^{1/2}}{24}\\ \qquad \qquad =\dfrac{\sqrt{145}}{288}\simeq 0.041811 \end{array} \end{aligned}$$

From now on, we will take \(c_2=1/2\) and \(c_3=3/4\). Thus, we have three coefficients \(\alpha , \beta _{22}, \beta _{32}\) to minimize the error coefficients and maximize the stability region.

The limit of the stability function of this method as z approaches infinity is given by

$$\begin{aligned} \lim _{z\rightarrow \infty } {\hat{R}}(z) = 1-{\hat{b}}^T({\hat{A}} + \varGamma )^{-1} \textbf{1} = \dfrac{a_0 + a_1\alpha -288 \alpha ^2 +96\alpha ^3}{96\alpha ^3}, \end{aligned}$$

with

$$\begin{aligned} \begin{array}{l} a_0= -(-3 + \beta _{22}) (-3 + \beta _{32}) (33 + 3 \beta _{22} + 4 \beta _{32}),\\ a_1= -6 (-45 + 12 \beta _{22} + \beta _{22}^2). \end{array} \end{aligned}$$

There are two values of \(\beta _{32}\) (as functions of \(\alpha \) and \(\beta _{22}\)) that make \({\widehat{R}}(\infty )=0\). We will select one of these values and use the coefficients \(\alpha \) and \(\beta _{22}\) to minimize the error coefficients and maximize the \(L(\theta )\)-stability angle.

The 2-norm of the error coefficients is given by

$$\begin{aligned} C_4=( C_{41}^2+ \cdots + C_{47}^2)^{1/2}. \end{aligned}$$

Minimizing this norm is equivalent to minimizing (the other coefficients are constant) \( C_{45}^2 + C_{46}^2 + C_{47}^2. \) A good compromise between large \(\theta \) and small \(C_4\) is obtained for \(\alpha = 0.54\), which yields \(\beta _{22} = -6.1\) and \(\beta _{32} = -2.75034\), resulting in a method, that we will name MSRKTASE3a, with stability angle and error coefficients given in Table 3. The coefficients of this method are given, with 32 digits, in the Appendix.

Alternatively, we can minimize the 2-norm of the coefficients of the leading term in the case that W is exactly the Jacobian matrix,

$$\begin{aligned} D_4=( D_{41}^2+ \cdots + D_{45}^2)^{1/2}. \end{aligned}$$

This is equivalent to minimize \(|D_{41}|= {\hat{b}}^T ( \varGamma +{\hat{A}}) ^3 \textbf{1} -1/24\) (the other coefficients are constant). This error coefficient can be vanished when

$$\begin{aligned} \beta _{22}=-3 - \dfrac{1}{3 \alpha ^2} + 8 \alpha . \end{aligned}$$

Now, the remaining coefficient \(\alpha \) can be used to minimizing \(\theta \) and \(C_4\). Unfortunately, when \(\alpha \) decreases \(C_4\) decreases but \(\theta \) increases. A good compromise between both quantities is obtained when \(\alpha = 0.56\), which yields \(\beta _{22} = 0.417075\) and \(\beta _{32} = -8.03347\), resulting in a method, that we will name MSRKTASE3b, with stability angle and error coefficients given in Table 3. The coefficients of this method, with 32 digits, are given in the Appendix.

In Fig.  1 we plot the boundary of the stability regions of the proposed methods. The method with minimal \(D_4\) is shown in red-dashed, the method with minimal \(C_4\) in red-solid, and the method from [4] in blue.

Fig. 1
figure 1

Boundary of the stability regions of the new methods (left). On the right, a zoom of the boundary near the origin

We can see from these results that the new method MSRKTASE3a, has similar stability properties to the method of order 3 in [4], but has about 35 times smaller error coefficients. The new method is expected to provide more accurate approximations than the one in [4]. The other method, MSRKTASE3b, also has much smaller error coefficients, although the stability angle is smaller. Nonetheless, its stability region is not significantly worse.

5 Numerical Experiments

To evaluate the performance of the new methods, we considered two test problems and integrated them using the two new methods, MSRKTASE3a and MSRKTASE3b. For comparison, we also used the method proposed in [4] (SRKTASE3) to demonstrate the improvements of these new modified singly-RKTASE methods over the previous singly-RKTASE methods. Additionally, to benchmark the performance of the new methods against other known Runge–Kutta methods for stiff problems, we integrated the problems with a Singly-Diagonally Implicit RK method of order 3 (SDIRK3), proposed in [9, 10], that uses s=4 stages, and also with the W-method ROW3b with s=3, order 2 but order 3 when \(W= F'\), proposed by Gonzalez-Pinto and coworkers in [11]. The norm of the coefficients of the leading term of the local error of these methods is given in Table 3. Since the SDIRK3 method is implicit, we used a simplified Newton method to solve the stage equations, approximating the Jacobian matrix with the matrix W used in the TASE operators.

For each method and problem, we computed the CPU time required for solving it and the 2-norm of the global error at the end of the integration interval. These data points allowed us to plot the global error (in logarithmic scale) against the CPU time, resulting in an efficiency plot.

Problem 1: (taken from [1] and [2]) The 1D diffusion of a scalar function \(y = y(x, t)\), with a time dependent source term

$$\begin{aligned} \dfrac{\partial y}{\partial t}= \dfrac{\partial ^2 y}{\partial x^2} + 0.1 \sin (t/50), \quad y(x, 0) = 1 - \cos (x)^{101}, \quad -\pi \le x \le \pi , \quad t\in [0,6]. \end{aligned}$$

The solution \(y = y(x, t)\) is assumed to be \(2\pi \)-periodic in x. For the spatial discretization, we used fourth-order centered difference schemes with a grid resolution of \(N = 512\), assuming periodic boundary conditions. The real part of the eigenvalues of the Jacobian matrix of the semi-discrete problem ranges from \(-3.5 \times 10^4\) to \(2.79 \times 10^{-5}\).

The efficiency plot of the methods for this problem is depicted in Fig. 2.

Fig. 2
figure 2

Efficiency plot for Diffusion problem

Since for this problem the operator \(I-\varDelta t\;\alpha W\) is constant, just one LU factorization is needed and the number of triangular linear systems solved is relevant for the total computational cost. The method MSRKTASE3b is as efficient as SDIRK3. MSRKTASE3b has higher computational cost per step (9 linear systems) than SDIRK3 (4 linear systems, since the problem is linear), but MSRKTASE3b has an error coefficient about ten times smaller than SDIRK3, resulting globally in similar efficiency. MSRKTASE3a is less efficient due to is larger error coefficient, and is clearly more efficient than SRKTASE3. On the other hand, ROW3b is the most efficient. This is due to its computational cost. ROW3b has an error coefficient larger than the ones of SDIRK3 and MSRKTASE3b but it requires only the solution of 3 linear systems per step.

Problem 2: The 1D Burgers’ equation written in the conservative form

$$\begin{aligned} \dfrac{\partial y}{\partial t}=0.1 \dfrac{\partial ^2 y}{\partial x^2} -\dfrac{\partial }{\partial x} \left( \dfrac{y}{2}\right) ^2, \quad y(x,0)= 1 -\cos (x)^{101}, \quad -\pi \le x \le \pi , \quad t\in [0,6]. \end{aligned}$$

The solution \(y=y(x,t)\) is assumed to be \(2\pi \)-periodic in x. For the spatial discretization, a fourth-order centered difference scheme with grid resolution of \(N=512\) has been taken.

The real part of the eigenvalues of the Jacobian matrix of the semi-discrete problem ranges from \(-3.5\times 10^{3}\) to \(2.6\times 10^{-6}\).

We integrated Burger’s problem using the four methods with three different options for the matrix W. Firstly, the Jacobian matrix is evaluated at every time step and \(W= \partial f(t_n, y_n)/\partial y \). With this option, the LU matrix factorization had to be computed at every step, increasing the computational cost. Secondly, the Jacobian matrix was evaluated only at the initial time step and \(W= \partial f(t_0, y_0)/\partial y \), requiring only one LU factorization and considerably reducing the CPU time. However, this could affect the accuracy in the case of TASE or W-methods, or the number of iterations required to solve the non linear system in the case of the SDIRK method. Finally, W was taken as the linear part of the semi-discrete differential equation (the matrix of the diffusion term). The computational cost is similar to the second option, but the accuracy could be lower. This option provided insight into the methods behaviour when W poorly approximates the Jacobian matrix.

Fig. 3
figure 3

Efficiency plot for Burger’s problem integrated with W the Jacobian matrix evaluated at every step (left) and W the Jacobian matrix evaluated just at the initial time step (right)

Figures 3 and 4 show the efficiency plots for Burger’s problem integrated with these three options. As shown in the Figures, evaluating the Jacobian matrix only at the initial step reduces the CPU time but increases the global error for the TASE-based and W-methods, where the local error depends on W and is smaller when it coincides with the Jacobian matrix. For the SDIRK3 method, the error does no vary significantly with different W, but solving the nonlinear system requires more iterations when the Jacobian matrix is less accurately approximated, thereby increasing the computational cost.

Fig. 4
figure 4

Efficiency plot for Burger’s problem integrated with W equal to the matrix of diffusion term

MSRKTASE3a exhibited stability issues with the two largest time steps when W is not the Jacobian matrix at every step, likely due to a high sensitivity of the stability function to parameter changes. MSRKTASE3b also showed stability problems with the largest time step size and the poorest approximation of the Jacobian matrix. Further research is being carried out in this regard.

Concerning the efficiency of the methods, when the Jacobian matrix is evaluated at every step, SDIRK3 is the most efficient, followed by ROW3b, MSRKTASE3b and then by MSRKTASE3a. SRKTASE3 is the least efficient due to its larger error coefficients. When the Jacobian matrix is evaluated only at the initial point, ROW3b is the most efficient, MSRKTASE3a and MSRKTASE3b are more efficient than the other two while SRKTASE3 remains the least efficient. Finally, when the matrix W is the matrix of the diffusion term, ROW3b is the most efficient for large time steps, MSRKTASE3a is more efficient for small time steps, where it has no stability problems, due to the relevance of the error coefficient C4. MSRKTASE3b remains more efficient than SDIRK3 and SRKTASE3 is also more efficient than SDIRK3 except for the smallest time step. The solution of the nonlinear equations makes SDIRK3 less efficient.

6 Conclusions

We presented a modification of the Singly-RKTASE methods for the numerical solution of stiff differential equations by taking different TASE operators for each stage of the RK method. We proved that these methods are equivalent to W-methods which enabled us to derive the order conditions when the TASE operators have order smaller than the order p of the Runge–Kutta scheme. For the case where \(p=3\) we obtained Modified Singly-RKTASE methods that significantly reduce the error coefficients compared to Singly-RKTASE methods, while maintaining stability.

Numerical experiments on both linear and nonlinear stiff systems demonstrate that the modified Singly-RKTASE methods provide significant improvements in accuracy and computational efficiency over the original Singly-RKTASE schemes, making them competitive with other methods such as Diagonally Implicit RK methods.