Free Access
Issue
A&A
Volume 655, November 2021
Article Number A87
Number of page(s) 12
Section Numerical methods and codes
DOI https://doi.org/10.1051/0004-6361/202141237
Published online 25 November 2021

© ESO 2021

1. Introduction

The theory of the generation and transfer of polarized radiation is of crucial importance for solar and stellar physics. Currently, one of the main challenges in the field is performing detailed numerical simulations of the transfer of polarized radiation in magnetized and dynamic atmospheres. In particular, it is essential to properly account for scattering processes in complex atomic models and multidimensional geometries, which lead to nonlinear and nonlocal multidimensional problems. In applied science, it is often profitable to convert nonlinear problems into linear ones, allowing the use of the powerful tools from numerical linear algebra, such as stationary or Krylov iterative methods. This is also the case in the field of numerical radiative transfer, where a fair number of problems of practical interest are accurately described through a linear system. For instance, the modeling of the transfer of resonance line polarization considering two-level (or two-term) atomic models can be linearized by assuming that the lower level (or term) is not polarized and that its population is known and fixed a priori (see, e.g., Belluzzi & Trujillo Bueno 2014; Sampoorna et al. 2017; Alsina Ballester et al. 2017).

In principle, linear problems can be solved in a single step with a direct method. However, this task can be highly inefficient or even unfeasible in practice, especially for large problems. Therefore, it is common to rely on efficient stationary iterative methods that, combined with suitable parallelization strategies, make the whole problem computationally tractable.

Trujillo Bueno & Landi Degl’Innocenti (1996) first applied the Richardson (a.k.a. fixed point iteration), Jacobi, and block-Jacobi iterative methods to the numerical transfer of polarized radiation1, while Trujillo Bueno & Manso Sainz (1999) applied Gauss-Seidel and successive over-relaxation (SOR) iterative methods to the same problem. In particular, the block-Jacobi iterative method is still frequently used in different scattering polarization problems (e.g., Belluzzi & Trujillo Bueno 2014; Alsina Ballester et al. 2017; Sampoorna et al. 2017; Alsina Ballester et al. 2018). By contrast, Gauss-Seidel and SOR iterative methods are not of common practice (Lambert et al. 2016), because they are nontrivial to parallelize. Despite of its slow convergence, the Richardson iterative method is still used in computationally complex transfer problems, where the other iterative methods are not exploitable (e.g., del Pino Alemán et al. 2020; Janett et al. 2021).

A rigorous investigation of the convergence conditions of such iterative methods applied to transfer problems of polarized radiation is still lacking. The specific stability requirements are problem-dependent and often difficult to determine. This paper gives a deeper analysis of these convergence conditions, analyzing their dependence on different design elements, such as the formal solver, the spectral, angular, and spatial numerical grids, and the damping factor.

Section 2 briefly introduces iterative methods, exposes their convergence conditions, and presents the most common stationary methods. Section 3 presents the benchmark analytical problem employed in this paper and exposes its discretization and algebraic formulation in the context of various iterative methods. Section 4 numerically analyzes the impact of different problem elements on the iterative solutions, focusing on the choice of the formal solver, on the various discrete grids, and on damping parameters. Finally, Sect. 5 provides remarks and conclusions, which are also generalized to more complex problems.

2. Iterative methods

We consider the linear system

A x = b , $$ \begin{aligned} A\mathbf{x }=\mathbf{b }, \end{aligned} $$(1)

where the nonsingular matrix A R N × N $ A\in\mathbb R^{N\times N} $ and the vector b R N $ {\mathbf{b}}\in \mathbb R^N $ are given, and the solution x R N $ {\mathbf{x}}\in \mathbb R^N $ is to be found. In absence of rounding errors, direct methods give the exact solution to the linear system (1) by a finite sequence of steps. They provide the solution

x = A 1 b , $$ \begin{aligned} \mathbf{x }=A^{-1}\mathbf{b }, \end{aligned} $$

using, for example, a suitable factorization of A, using Gaussian elimination. Due to their high computational complexity and storage costs, direct methods are often prohibitive in practice, especially for large problems. For this reason, iterative methods are usually preferred for the solution of large linear systems. These methods are convenient in terms of parallel implementation and therefore particularly suitable for today’s high performance computing systems.

Iterative methods start with an initial guess x 0 R N $ {\mathbf{x}}^0\in\mathbb R^N $ and generate a sequence of improving approximate solutions x 1 , x 2 , , x n , x n + 1 R N $ {\mathbf{x}}^1,{\mathbf{x}}^2,\dots,{\mathbf{x}}^n,{\mathbf{x}}^{n+1}\in\mathbb R^N $ given by

x 1 = ϕ 0 ( x 0 ; A , b ) , x 2 = ϕ 1 ( x 0 , x 1 ; A , b ) , x n + 1 = ϕ n ( x 0 , x 1 , , x n ; A , b ) . $$ \begin{aligned}&\mathbf{x }^1 = \phi _0(\mathbf{x }^0;A,\mathbf{b }),\\&\mathbf{x }^2 = \phi _1(\mathbf{x }^0,\mathbf{x }^1;A,\mathbf{b }),\\&\quad \ \vdots \\&\mathbf{x }^{n+1} = \phi _n(\mathbf{x }^0,\mathbf{x }^1,\dots ,\mathbf{x }^n;A,\mathbf{b }). \end{aligned} $$

A one-step iterative method computes the approximate solution xn+1 by considering only the previous approximate solution xn. Hence, the transition from xn to xn+1 does not rely on the previous history of the iterative process. Moreover, a stationary iterative method generates the sequence of approximate solutions by the recursive application of the same operator ϕ. Therefore, stationary one-step iterative methods can be expressed in the form

x n + 1 = ϕ ( x n ; A , b ) . $$ \begin{aligned} \mathbf{x }^{n+1}=\phi (\mathbf{x }^n;A,\mathbf{b }). \end{aligned} $$

If the operator ϕ is also linear, the linear stationary one-step iterative method can be expressed in the form

x n + 1 = G x n + g , $$ \begin{aligned} \mathbf{x }^{n+1}=G\mathbf{x }^n+\mathbf g , \end{aligned} $$(2)

where neither the iteration matrix G R N × N $ G\in \mathbb R^{N\times N} $ nor the vector g depend upon the iteration n. Hence, a linear stationary one-step iterative method starts from an initial guess x0 and repeatedly invokes the recursive formula (2) until a chosen termination condition matches (cf. Sect. 2.4).

In order to study and design linear stationary one-step iterative methods, it is convenient to consider the linear system (1) in the form

x = G x + g . $$ \begin{aligned} \mathbf{x }=G\mathbf{x }+\mathbf g . \end{aligned} $$(3)

We recall that multistep iterative methods have been also used in the radiative transfer context. Relaying on multiple previous iterates, multistep iterative methods can accelerate the convergence of one-step iterations. In this regard, we mention Ng (e.g., Olson et al. 1986), ORTHOMIN (e.g., Klein et al. 1989), and state-of-the-art Krylov methods. In particular, Krylov acceleration strategies are the subject of the second part of this paper series (Benedusi et al. 2021).

2.1. Convergence conditions

An iterative method is convergent if the approximate solution xn converges to the exact solution x as the iteration process progresses, that is, if xnx for n → ∞. This implies that the error, defined as

e n = x n x , $$ \begin{aligned} \mathbf e ^n=\mathbf{x }^n-\mathbf{x }, \end{aligned} $$

must approach zero as n increases. Subtracting Eq. (3) from Eq. (2), one obtains

e n + 1 = G e n , $$ \begin{aligned} \mathbf e ^{n+1}=G\mathbf e ^n, \end{aligned} $$

indicating that the convergence of the iterative method depends entirely on the iteration matrix G. In fact, the iterative method (2) is convergent for any initial guess x0 if and only if

ρ ( G ) < 1 , $$ \begin{aligned} \rho (G) < 1, \end{aligned} $$(4)

where ρ(G) is the spectral radius of G (see, e.g., Hackbusch 2016). The quantity ρ(G) is also known as the convergence rate, because it describes the speed of convergence of an iterative process. When employing an iterative method, a rigorous convergence analysis based on the condition (4) is advisable. However, since an analytical expression of ρ(G) is often hard to obtain in practice, heuristic-based analyses are commonly employed to study how the convergence of iterative methods depends on problem and discretization parameters (Hackbusch 2016).

2.2. Examples

Linear stationary one-step iterative methods for the solution of (1) usually arise from the splitting

A = P R , $$ \begin{aligned} A=P-R, \end{aligned} $$(5)

with P nonsingular. Rewriting Eq. (1) in the form of Eq. (3) and introducing iteration indices as in Eq. (2), one obtains

x n + 1 = P 1 R x n + P 1 b , $$ \begin{aligned} \mathbf{x }^{n+1}=P^{-1}R\mathbf{x }^n+P^{-1}\mathbf{b }, \end{aligned} $$(6)

with G = P−1R = P−1(PA) = IdP−1A and g = P−1b. Crucially, P has to be easier to invert then A, while being a good approximation of it to ensure fast convergence.

Sometimes, it is convenient to express iterative methods in terms of the correction

δ x n = x n + 1 x n . $$ \begin{aligned} \boldsymbol{\delta } \mathbf{x }^n = \mathbf{x }^{n+1}-\mathbf{x }^n. \end{aligned} $$

In the correction form, the iterative method (6) reads

x n + 1 = ( I d P 1 A ) x n + P 1 b = x n + P 1 ( b A x n ) , $$ \begin{aligned} \mathbf{x }^{n+1}=(Id-P^{-1}A)\mathbf{x }^n+P^{-1}\mathbf{b }= \mathbf{x }^n + P^{-1}(\mathbf{b }- A\mathbf{x }^n), \end{aligned} $$

or simply

x n + 1 = x n + δ x n , $$ \begin{aligned} \mathbf{x }^{n+1}=\mathbf{x }^n + {\boldsymbol{\delta }} \mathbf{x }^n, \end{aligned} $$(7)

where the correction is given by δxn = P−1rn and rn = bAxn is the iteration residual.

Many iterative methods are defined through the following decomposition of the matrix A

A = D + L + U , $$ \begin{aligned} A=D+L+U, \end{aligned} $$

where D is the diagonal of A, while L and U are its strict lower and strict upper triangular parts, respectively. Different choices of the matrix P yield different iterative methods:

P = I d , Richardson method , P = D , Jacobi method , P = blockdiag ( A 1 , , A m ) , block-Jacobi method , P = D + L or P = D + U , Gauss-Seidel method , $$ \begin{aligned} P&=Id,&\text{ Richardson} \text{ method},\\ P&=D,&\text{ Jacobi} \text{ method},\\ P&=\text{ blockdiag}(A_1,\ldots ,A_m),&\text{ block-Jacobi} \text{ method},\\ P&=D+L \text{ or} P=D+U,&\text{ Gauss-Seidel} \text{ method}, \end{aligned} $$

with A1, …, Am being m square diagonal blocks of A. All these choices of P allow for a convenient, and possibly parallel, computation of P−1. We note that the iterative method corresponding to P = A always converges in one iteration, resulting in a direct method.

In another perspective, known as preconditioning, all these methods correspond to a Richardson iteration applied to the preconditioned system

P 1 A x = P 1 b , $$ \begin{aligned} P^{-1}A\mathbf{x }=P^{-1}\mathbf{b }, \end{aligned} $$(8)

which is equivalent to the linear system (1). The nonsingular matrix P is known as preconditioner and is usually chosen to reduce the condition number of P−1A with respect to the one of A, such that the preconditioned system is “easier” to solve with respect to the original one.

Multiplying the preconditioner P (or part of it) by a positive scalar, leads to damped iterative schemes, which are discussed in the following section.

2.3. Damped iterative methods

The classical idea of damping (or relaxation) consists in scaling the correction δxn in Eq. (7) by a damping factor ω R > 0 $ \omega\in\mathbb R_{ > 0} $, namely

x n + 1 = x n + ω δ x n = x n + ω P 1 r n . $$ \begin{aligned} \mathbf{x }^{n+1}=\mathbf{x }^n + \omega \boldsymbol{\delta } \mathbf{x }^n = \mathbf{x }^n + \omega P^{-1}\mathbf r ^n. \end{aligned} $$(9)

The term damped properly holds only for 0 < ω < 1, a.k.a. under-relaxation. In fact, the choice ω = 1 yields the original method, while ω > 1 gives an extrapolated version, a.k.a. over-relaxation. For simplicity, the term damped will be used for all choices of ω. Damped iterations can also be read as

x n + 1 = x n ω x n + ω x n + ω δ x n + 1 = ( 1 ω ) x n + ω x n + 1 , $$ \begin{aligned} \mathbf{x }^{n+1}=\mathbf{x }^n-\omega \mathbf{x }^n+\omega \mathbf{x }^n+\omega \boldsymbol{\delta } \mathbf{x }^{n+1}=(1-\omega )\mathbf{x }^n+\omega \tilde{\mathbf{x }}^{n+1}, \end{aligned} $$

that is, the weighted mean of the old iterate xn and one step of the original method x n + 1 = x n + δ x n $ \tilde{{\mathbf{x}}}^{n+1} = {\mathbf{x}}^n+{\boldsymbol{\delta}} {\mathbf{x}}^n $.

In order to guarantee the convergence of damped methods, it is possible to obtain bounds on ω as well as optimal choices in terms of convergence rate. In particular, the preconditioned iteration (9) is convergent if and only if

ω < 2 Re ( λ i ) | λ i | 2 , for all i = 1 , , N , $$ \begin{aligned} \omega < \frac{2\text{ Re}(\lambda _i)}{|\lambda _i|^2}, \quad \mathrm{for\ all} \quad i = 1,\ldots ,N, \end{aligned} $$(10)

λi being the ith eigenvalue of the matrix P−1A. If P−1A has real positive eigenvalues, the optimal convergence rate for the iterative methods presented in Sect. 2.2 is attained for

ω opt = 2 λ min + λ max , $$ \begin{aligned} \omega _{\text{opt}}=\frac{2}{\lambda _{\min }+\lambda _{\max }}, \end{aligned} $$(11)

where λmin and λmax are the minimum and maximum eigenvalues of P−1A. The SOR method is a damped variant of the Gauss-Seidel method, with P = ω−1D + L. If A is consistently ordered, GJac = IdD−1A has real eigenvalues, and ρ(GJac) < 1, the optimal convergence rate for SOR is attained for

ω sor = 1 + ( ρ ( G Jac ) 1 + 1 ρ ( G Jac ) 2 ) 2 . $$ \begin{aligned} \omega _{\text{sor}}=1+\left(\frac{\rho (G_{\text{Jac}})}{1+\sqrt{1-\rho (G_{\text{Jac}})^2}}\right)^2. \end{aligned} $$(12)

A deeper and comprehensive analysis of damped stationary iterative methods is given, for instance, by Hageman & Young (1981), Saad (2003), Quarteroni et al. (2010), Hackbusch (2016).

2.4. Termination condition

Typically, iterative solution strategies terminate at the nth step if the norm of the relative residual is smaller then a desired tolerance, that is, if ‖rn2/‖b2 < tol. However, this stopping criteria becomes less and less reliable as the condition number of A, denoted with κ(A), increases. In fact, the following relation between the relative error and the relative residual holds:

x x n 2 / x 2 κ ( A ) r n 2 / b 2 . $$ \begin{aligned} \Vert \mathbf{x }-\mathbf{x }^n\Vert _2/\Vert \mathbf{x }\Vert _2\le \kappa (A)\Vert \mathbf{r }^n\Vert _2/\Vert \mathbf{b }\Vert _2. \end{aligned} $$

Since by construction κ(P−1A) < κ(A), the same termination condition with respect to the preconditioned system (8), namely, ‖P−1rn2/‖P−1b2 < tol, gives a better approximation of the true relative error.

3. Benchmark problem

In this section, we present the continuous formulation of a linear benchmark transfer problem for polarized radiation. Secondly, we consider its discretization and derive its algebraic formulation. Finally, we remark how the stationary iterative methods presented in Sects. 2.2 and 2.3 are applied in this context.

The problem is formulated within the framework of the theoretical approach described in Landi Degl’Innocenti & Landolfi (2004). Within this theory, a scattering process is treated as a succession of independent absorption and reemission processes. This is the so-called limit of complete frequency redistribution (CRD), in which no frequency correlations between the incoming and the outgoing photons in a scattering process are taken into account. We consider a two-level atom with an unpolarized lower level. Since stimulated emission is completely negligible in solar applications, it is not considered. For the sake of simplicity, the contribution of continuum processes is neglected, as well as that of magnetic and bulk velocity fields. We note, however, that the methodology presented in this section can also be applied to a wider range of settings, such as the unpolarized case, two-term atomic models including continuum contributions, atmospheric models with arbitrary magnetic and bulk velocity fields, and theoretical frameworks accounting for partial frequency redistribution (PRD) effects.

3.1. Continuous problem

The physical quantities entering the radiative transfer problem are, in general, functions of the spatial point r, the frequency ν, and the propagation direction Ω of the radiation ray under consideration. In the absence of polarization phenomena, the specific intensity and the atomic level population fully describe the radiation field and the atomic system, respectively. When polarization is taken into account, a more detailed description is instead required. The radiation field is fully described by the four Stokes parameters

I i ( r , Ω , ν ) , $$ \begin{aligned} I_i(\boldsymbol{r},\mathbf{\Omega },\nu ), \end{aligned} $$

with i = 1, …, 4, standing for Stokes I, Q, U, and V, respectively. The nth energy level of the atomic system, with total angular momentum Jn, is fully described by the multipolar components of the density matrix, a.k.a. spherical statistical tensors,

ρ Q , n K ( r ) , $$ \begin{aligned} \rho _{Q,n}^{K}(\boldsymbol{r}), \end{aligned} $$

with K = 0, …, 2Jn, and Q = −K, …, K. We recall that the spherical statistical tensors are in general complex quantities, and that the 0-rank component, ρ 0,n 0 $ \rho_{0,n}^0 $, is proportional to the population of the nth level (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 3). A two-level atomic system at position r is thus described by the following quantities

ρ Q , K ( r ) , and ρ Q , u K ( r ) , $$ \begin{aligned} \rho _{Q,\ell }^{K}(\boldsymbol{r}), \; \ \mathrm{and }\; \ \rho _{Q,u}^{K}(\boldsymbol{r}), \end{aligned} $$

with the subscripts and u indicating the lower and upper levels, respectively. The assumption of unpolarized lower level2 implies that

ρ Q , K ( r ) = ρ 0 , 0 ( r ) δ K 0 δ Q 0 , $$ \begin{aligned} \rho ^K_{Q,\ell }(\boldsymbol{r}) = \rho ^0_{0,\ell }(\boldsymbol{r}) \delta _{K0} \delta _{Q0}, \end{aligned} $$

where δij is the Kronecker delta.

In the absence of magnetic fields, the statistical equilibrium equations of a two-level atom with an unpolarized lower level have the following analytical solution (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 14)3:

ρ Q , u K ( r ) ρ 0 , 0 ( r ) = 2 J u + 1 2 J + 1 c 2 2 h ν 0 3 [ a J u J ( K , Q ) ( r ) J ¯ Q K ( r ) + ϵ ( r ) W ( r ) δ K 0 δ Q 0 ] , $$ \begin{aligned}&\frac{ \rho ^K_{Q,u}(\boldsymbol{r})}{\rho ^0_{0,\ell }(\boldsymbol{r})} = \sqrt{\frac{2J_u + 1}{2J_\ell + 1}} \frac{c^2}{2 h \nu _0^3} \left[ a_{J_u J_\ell }^{(K,Q)}(\boldsymbol{r})\bar{J}^K_{-Q}(\boldsymbol{r}) + \epsilon (\boldsymbol{r})W(\boldsymbol{r})\delta _{K0} \delta _{Q0} \right], \end{aligned} $$(13)

where c and h have their usual meaning of light speed and Planck constant, respectively, ϵ is the so-called thermalization parameter, W is the Planck function in the Wien limit at the line-center frequency ν0, and

a J u J ( K , Q ) ( r ) = ( 1 ) Q ( 1 ϵ ( r ) ) w J u J ( K ) 1 + δ u ( K ) ( r ) ( 1 ϵ ( r ) ) . $$ \begin{aligned} a_{J_u J_\ell }^{(K,Q)}(\boldsymbol{r})=(-1)^Q\frac{(1-\epsilon (\boldsymbol{r})) { w}^{(K)}_{J_u J_\ell }}{1 + \delta ^{(K)}_u(\boldsymbol{r}) (1-\epsilon (\boldsymbol{r}))}. \end{aligned} $$

The coefficient w J u J l (K) $ {\it w}^{(K)}_{J_u J_\ell} $ is the polarizability factor4, δ u (K) $ \delta^{(K)}_u $ describes the depolarizing rate of the upper level due to elastic collisions, and the frequency-averaged radiation field tensor is given by

J ¯ Q K ( r ) = d ν ϕ ( r , ν ) 1 4 π d Ω i = 1 4 T Q , i K ( Ω ) I i ( r , Ω , ν ) , $$ \begin{aligned} \bar{J}^K_Q(\boldsymbol{r})= \int \mathrm{d} \nu \,\phi (\boldsymbol{r},\nu ) \frac{1}{4 \pi } \oint \mathrm{d} \mathbf {\Omega} \sum _{i=1}^4\mathcal{T} _{Q,i}^K(\mathbf {\Omega} ) I_i(\boldsymbol{r},\mathbf {\Omega} ,\nu ), \end{aligned} $$(14)

where T Q,i K ( Ω ) $ \mathcal{T}^K_{Q,i}(\mathbf{\Omega}) $ is the polarization tensor (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 5) and ϕ is the absorption profile. By definition, the maximum rank of the radiation field tensor J ¯ Q K $ \bar{J}^K_Q $ is K = 2 (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 5). Consequently, the only nonzero spherical statistical tensors of the upper level ρ Q,u K $ \rho^K_{Q,u} $ are those with K ≤ 2. Exploiting the conjugation properties of the spherical statistical tensors (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 3), at each position r, the considered two-level atomic system with an unpolarized lower level is thus described by 9 real quantities.

The propagation of the Stokes parameters at frequency ν along Ω is described by the following differential equation

d d s I i ( r , Ω , ν ) = j = 1 4 K ij ( r , Ω , ν ) I j ( r , Ω , ν ) + ε i ( r , Ω , ν ) , $$ \begin{aligned} \frac{\mathrm{d} }{\mathrm{d} s} I_i(\boldsymbol{r},\boldsymbol{\Omega },\nu ) = - \sum _{j=1}^4 K_{ij}(\boldsymbol{r},\boldsymbol{\Omega },\nu ) I_j(\boldsymbol{r},\boldsymbol{\Omega },\nu ) + \varepsilon _i(\boldsymbol{r},\boldsymbol{\Omega },\nu ), \end{aligned} $$

where s is the spatial coordinate along the direction Ω, εi is the emission coefficient of the ith Stokes parameter, and the entries Kij form the 4 × 4 propagation matrix.

In the absence of a magnetic field and lower level polarization, the propagation matrix is diagonal with diagonal element η. The transfer equations for the four Stokes parameters are consequently decoupled and read

d d s I i ( r , Ω , ν ) = η ( r , ν ) [ I i ( r , Ω , ν ) S i ( r , Ω ) ] . $$ \begin{aligned} \frac{\mathrm{d} }{\mathrm{d} s} I_i(\boldsymbol{r},\boldsymbol{\Omega },\nu ) = - \eta (\boldsymbol{r},\nu ) \left[I_i(\boldsymbol{r},\boldsymbol{\Omega },\nu ) - S_{\!i}(\boldsymbol{r},\boldsymbol{\Omega }) \right]. \end{aligned} $$(15)

We note that η does not depend on Ω because bulk velocity fields are neglected, while the source function Si does not depend on ν because of the CRD assumption. Considering dipole scattering and neglecting the continuum contribution, the source function for the ith Stokes parameter is given by (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 14)

S i ( r , Ω ) = ε i ( r , Ω , ν ) η ( r , ν ) = KQ w J u J ( K ) T Q , i K ( Ω ) σ Q K ( r ) , $$ \begin{aligned} S_{\!i}(\boldsymbol{r},\boldsymbol{\Omega })= \frac{\varepsilon _i(\boldsymbol{r},\boldsymbol{\Omega },\nu )}{\eta (\boldsymbol{r},\nu )} =\sum _{KQ} { w}^{(K)}_{J_u J_\ell }\mathcal{T} ^K_{Q,i}(\boldsymbol{\Omega }) \sigma ^K_Q(\boldsymbol{r}), \end{aligned} $$(16)

with

σ Q K ( r ) = 2 h ν 0 3 c 2 2 J + 1 2 J u + 1 ρ Q , u K ( r ) ρ 0 , 0 ( r ) . $$ \begin{aligned} \sigma ^K_Q(\boldsymbol{r}) = \frac{2 h \nu _0^3}{c^2} \sqrt{\frac{2J_\ell + 1}{2J_u + 1}} \frac{ \rho ^K_{Q,u}(\boldsymbol{r})}{\rho ^0_{0,\ell }(\boldsymbol{r})}. \end{aligned} $$(17)

Using Eq. (13), the quantities σ Q K $ \sigma^K_Q $, which can be interpreted as the multipolar components of the source function, are given by

σ Q K ( r ) = a J u J ( K , Q ) ( r ) J ¯ Q K ( r ) + ϵ ( r ) W ( r ) δ K 0 δ Q 0 . $$ \begin{aligned} \sigma ^K_Q(\boldsymbol{r}) = a_{J_u J_\ell }^{(K,Q)}(\boldsymbol{r}) \bar{J}^K_{-Q}(\boldsymbol{r}) + \epsilon (\boldsymbol{r}) W(\boldsymbol{r}) \delta _{K0} \delta _{Q0}. \end{aligned} $$(18)

3.2. Discretization and algebraic formulation

In order to replace the continuous problem by a system of algebraic equations, we discretize the continuous variables ν, r, and Ω using, respectively, grids with Nν, Ns and NΩ nodes. The quantities of the problem are now evaluated at the nodes at { r k } k = 1 N s $ \{{\mathbf{r}}_k\}_{k=1}^{N_s} $, { Ω m } m=1 N Ω $ \{\mathbf{\Omega}_m\}_{m=1}^{N_\Omega} $, and { ν p } p=1 N ν $ \{\nu_p\}_{p=1}^{N_{\nu}} $ only.

For k = 1, …, Ns, the quantity σ Q K ( r k ) $ \sigma^K_Q({\mathbf{r}}_k) $ is computed through the discrete versions of Eqs. (18) and (14), namely,

σ Q K ( r k ) = i = 1 4 m = 1 N Ω p = 1 N ν J K Q , i ( r k , Ω m , ν p ) I i ( r k , Ω m , ν p ) + c Q K ( r k ) , $$ \begin{aligned} \sigma ^K_Q(\mathbf{r }_k)\!=\!\sum _{i=1}^4\sum _{m=1}^{N_\Omega }\sum _{p=1}^{N_\nu } J_{KQ,i}(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p)I_i(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p)+ c_Q^K(\mathbf{r }_k), \end{aligned} $$(19)

where JKQ,i(rkm,νp) depends on the choice of spectral and angular quadratures used in Eq. (14) (cf. Eq. (A.3)), while c Q K ( r k ) = ϵ ( r k ) W ( r k ) δ K 0 δ Q 0 $ c_Q^K({\mathbf{r}}_k)=\epsilon({\mathbf{r}}_k)W({\mathbf{r}}_k)\delta_{K0}\delta_{Q0} $.

Similarly, the discrete version of Eq. (16) reads

S i ( r k , Ω m ) = KQ T i , K Q ( Ω m ) σ Q K ( r k ) , $$ \begin{aligned} S_{\!i}(\mathbf{r }_k,\mathbf{\Omega }_m)=\sum _{KQ}T_{i,KQ}(\mathbf{\Omega }_m)\sigma ^K_Q(\mathbf{r }_k), \end{aligned} $$(20)

with T i,KQ ( Ω m )= w J u J l (K) T Q,i K ( Ω m ) $ T_{i,KQ}({\bf\Omega}_m)={\it w}^{(K)}_{J_u J_\ell} \mathcal{T}^K_{Q,i}({\bf\Omega}_m) $. Finally, provided initial conditions, the transfer equation (15) can be solved along each propagation direction Ωm and at each frequency νp to provide the Stokes component Ii(rkm,νp) for all k. This numerical step can be expressed as

I i ( r k , Ω m , ν p ) = l = 1 N s Λ i ( r k , r l , Ω m , ν p ) S i ( r l , Ω m ) + t i ( r k , Ω m , ν p ) , $$ \begin{aligned}&I_i(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p)=\sum _{l=1}^{N_s}\Lambda _i(\mathbf{r }_k,\mathbf{r }_l,\mathbf{\Omega }_m,\nu _p)S_{\!i}({\mathbf{r }_l,\mathbf {\Omega} }_m)\nonumber \\&\qquad \qquad \qquad \ +t_i(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p), \end{aligned} $$(21)

where ti(rkm,νp) represents the radiation transmitted from the boundaries and the coefficients Λi(rk,rl,Ωm,νp) encode the numerical method used to propagate the ith Stokes parameter along the ray Ωm (cf. Eq. (C.9)). For notational simplicity, the explicit dependence on variables and indexes is omitted, and Eqs. (19)–(21) are expressed, respectively, in the following compact matrix form

σ = J I + c , with J R 9 N s × 4 N s N Ω N ν , $$ \begin{aligned}&\boldsymbol{\sigma }=J\mathbf I +\mathbf c ,\quad \mathrm{with}\ \ J\in \mathbb{R} ^{9 N_s\times 4 N_s N_\Omega N_\nu },\end{aligned} $$(22)

S = T σ , with T R 4 N s N Ω × 9 N s , $$ \begin{aligned}&\mathbf S =T\boldsymbol{\sigma },\qquad \mathrm{with}\ \ T\in \mathbb{R} ^{4 N_s N_\Omega \times 9 N_s},\end{aligned} $$(23)

I = Λ S + t , with Λ R 4 N s N Ω N ν × 4 N s N Ω , $$ \begin{aligned}&\mathbf I =\Lambda \mathbf S +\mathbf t ,\quad \mathrm{with}\ \ \Lambda \in \mathbb{R} ^{4 N_s N_\Omega N_\nu \times 4 N_s N_\Omega }, \end{aligned} $$(24)

where the vector I R 4 N s N Ω N ν $ \mathbf{I}\in\mathbb R^{4 N_s N_\Omega N_\nu} $ collects the discretized Stokes parameters, S R 4 N s N Ω $ \mathbf{S}\in\mathbb R^{4 N_s N_\Omega } $ the discretized source function, and σ R 9 N s $ {\boldsymbol{\sigma}}\in\mathbb R^{9 N_s} $ its discretized multipolar components. The explicit expressions of vectors and matrices appearing in Eqs. (22)–(24) are given in Appendices AC for a one-dimensional (1D) spatial grid. In general, the entries of the matrix J depend on the type of numerical integration used in Eq. (14), while the entries of T depend on the coefficients of Eq. (16). Moreover, the entries of Λ depend on the numerical method used to solve the transfer equation (15) (a.k.a. formal solver), on the spatial grid and, eventually, on the numerical conversion to the optical depth scale (e.g., Janett & Paganini 2018).

By choosing σ as the unknown vector, Eqs. (22)–(24) can then be combined into a single discrete problem, namely,

( I d J Λ T ) σ = J t + c , $$ \begin{aligned} (Id-J\Lambda T)\boldsymbol{\sigma }=J\mathbf t +\mathbf c , \end{aligned} $$(25)

with IdJΛT being a square matrix of size 9Ns.

The operator Λ and the vector t depend on the absorption coefficient η, which linearly depends on ρ 0,l 0 $ \rho^0_{0,\ell} $, which in turn enters the definition of the unknown σ. The discrete problem (25) is consequently nonlinear. However, the nonlinear problem (25) becomes linear if we further assume either that (i) the transfer equation (15) is solved by introducing the optical depth scale defined by

d τ ( ν ) = η ( r , ν ) d s , $$ \begin{aligned} \mathrm{d} \tau (\nu ) = -\eta (\boldsymbol{r},\nu )\mathrm{d} s, \end{aligned} $$

and the optical depth grid is fixed a priori, or that (ii) the density matrix element ρ 0,l 0 $ \rho^0_{0,\ell} $ (or equivalently the lower level atomic population) is fixed a priori and, consequently, the absorption coefficient η becomes a constant of the problem. We note that in both cases the operator Λ and the vector t become independent of the unknown σ.

In the literature, the so-called Λ-operator usually describes the mapping of the discretized source function vector S either to the specific intensity vector I or to the corresponding mean radiation field vector (e.g., Olson et al. 1986; Rybicki & Hummer 1991; Paletou & Auer 1995). This operator was introduced to highlight the linear nature of this mapping inside the whole nonlinear radiative transfer problem. The traditional Λ-operator corresponds either to Eq. (24) or to the combined action of Eq. (24) and the integral (14).

We note that the linearized system (25) has the same form as Eq. (1) with A = IdJΛT, x = σ, and b = Jt + c. The matrix A can be assembled by constructing J, Λ, and T, computing the products JΛT, and finally calculating IdJΛT. Alternatively, the matrix JΛT can be assembled column-by-column, constructing the jth column by applying Eqs. (19)–(21) (setting c q K $ c_q^K $ and ti to zero) to a point-like multipolar component of the source function σi = δij, for i, j = 1, …, 9Ns.

Applying the iterative methods exposed in Sect. 2.2 to Eq. (25), the following iteration matrix is obtained

G = I d P 1 ( I d J Λ T ) , $$ \begin{aligned} G = Id- P^{-1}(Id-J\Lambda T), \end{aligned} $$(26)

P being an arbitrary preconditioner. Accordingly, the nth iteration in the correction form reads

σ n + 1 = σ n + δ σ n , $$ \begin{aligned} \boldsymbol{\sigma }^{n+1}=\boldsymbol{\sigma }^n+\boldsymbol{\delta }\boldsymbol{\sigma }^n, \end{aligned} $$

where

δ σ n = P 1 ( J Λ T σ n σ n + J t + c ) . $$ \begin{aligned} \boldsymbol{\delta }\boldsymbol{\sigma }^n= P^{-1}\left(J\Lambda T \boldsymbol{\sigma }^n-\boldsymbol{\sigma }^n +J\mathbf t +\mathbf c \right). \end{aligned} $$

Sections 3.33.6 describe stationary iterative methods in the context of radiative transfer problems.

3.3. Richardson method

The form of the linear system (25) suggests the operator splitting Eq. (5) with P = Id and R = JΛT. The iterative method based on this splitting, that is,

σ n + 1 = J Λ T σ n + J t + c , $$ \begin{aligned} \boldsymbol{\sigma }^{n+1}=J\Lambda T \boldsymbol{\sigma }^n+J\mathbf t +\mathbf c , \end{aligned} $$(27)

is an example of the Richardson method, which is commonly known as Λ-iteration in the context of linear radiative transfer problems. Equation (27) describes the following iterative process: (i) given an initial estimate of σ, calculate S at each discrete position and direction through Eq. (16); (ii) compute the numerical solution of the transfer equation (15) to obtain I at each discrete position, frequency, and direction; (iii) integrate I at each position according to Eq. (14) to obtain the frequency averaged radiation field tensor, which is used to update σ using Eq. (18).

The damped Richardson method is simply obtained by multiplying the correction δσn of the Richardson method by a damping factor ω.

3.4. Jacobi method

Considering Eq. (25), an usual operator splitting Eq. (5) is given by choosing the preconditioner P to be diagonal, with the same diagonal of IdJΛT, resulting in a minimal cost for the application and computation of P−1, possibly in parallel. This splitting yields the Jacobi method, while its damped version is obtained by multiplying the correction δσn by a damping factor ω.

Trujillo Bueno & Manso Sainz (1999) applied the Jacobi method when dealing with CRD scattering line polarization linear problems. In this context, the Jacobi method is sometimes termed as ALI method or local ALI method (see Hubeny 2003). However, an increasing problem size results in a lower impact of the Jacobi acceleration. For instance, when PRD effects are included, the size of the resulting linear system scales with NsNν under the angle-averaged approximation (see, e.g., Alsina Ballester et al. 2017), while it scales with NsNνNΩ in the general angle-dependent treatment (see, e.g., Janett et al. 2021). In both cases, the impact of the Jacobi acceleration is barely perceptible.

3.5. Block-Jacobi method

In radiative transfer problems, linear systems often have a natural block structure, which offers different possibilities in defining the blocks. Appendices AC explicitly illustrate the structure of the matrices that build up the linear system (25) for the 1D case. Considering Eq. (25), the preconditioner P can be chosen as the block-diagonal of IdJΛT. This choice yields the block-Jacobi method, while its damped version is simply obtained by multiplying the correction δσn by a damping factor ω. We remark that, within this method, the computation of P−1 requires solving the multiple smaller systems corresponding to the diagonal blocks.

Trujillo Bueno & Manso Sainz (1999) first applied the block-Jacobi method to the CRD radiative transfer problem, choosing square blocks that account for the coupling of the different multipolar components of the source function. Moreover, the block-Jacobi method is a common choice when dealing with PRD radiative transfer problems under the angle-averaged approximation (e.g., Belluzzi & Trujillo Bueno 2014; Alsina Ballester et al. 2017; Sampoorna et al. 2017; Alsina Ballester et al. 2018). In this case, the preconditioner is usually built by square blocks of size Nν that account for the coupling of all the frequencies (see the frequency-by-frequency method of Paletou & Auer 1995). In such applications, the block-Jacobi method is referred to as ALI, Jacobi, or Jacobi-based method. Janett et al. (2021) first applied the damped block-Jacobi method to the same problem.

3.6. Gauss-Seidel method

Another choice of the operator splitting Eq. (5) is given by choosing the preconditioner P as the lower (or upper) triangular part of IdJΛT. This splitting yields the Gauss-Seidel method. The SOR method is simply obtained by modifying the correction δσn with a suitable damping factor ω as explained in Sect. 2.3. Trujillo Bueno & Manso Sainz (1999) first proposed and applied the Gauss-Seidel and SOR methods to CRD scattering line polarization linear problems, while Sampoorna & Trujillo Bueno (2010) generalized these methods to the PRD case. Trujillo Bueno & Fabiani Bendicho (1995) confirmed with demonstrative results that the optimal parameter ωsor given by Eq. (12) is indeed effective in the context of numerical radiative transfer. However, Gauss-Seidel and SOR methods are not common in radiative transfer applications (Lambert et al. 2016), mainly because they have limited and nontrivial parallelization capabilities (see, e.g., the red-black Gauss-Seidel algorithm Saad 2003).

4. Numerical analysis

It is worth mentioning that realistic radiative transfer problems can be substantially more complex than the one considered here. However, we present a suitable benchmark to better understand how the convergence of various iterative methods depends on different design elements, such as the choice of the formal solver, the discretization of the problem, or the use of damping factors. The conclusions could be then generalized to more complex problems.

In the absence of magnetic and bulk velocity fields, 1D plane-parallel atmospheres satisfy cylindrical symmetry with respect to the vertical. Due to the symmetry of the problem, the only nonzero radiation field tensor components are J ¯ 0 0 $ \bar{J}^0_0 $ and J ¯ 0 2 $ \bar{J}^2_0 $ and, from Eq. (18), the only nonzero multipolar components of the source function are σ 0 0 $ \sigma^0_0 $ and σ 0 2 $ \sigma^2_0 $. Setting the angle γ = 0 in the explicit expression of the polarization tensor T 0,i 2 $ \mathcal{T}^2_{0,i} $ entering Eq. (16), the only nonzero source functions are S1 and S2 (i.e., SI and SQ). Consequently, the only nonzero Stokes parameters are I1 and I2 (i.e., I and Q). Moreover, the angular dependence of the problem variables is fully described by the inclination θ ∈ [0, π] with respect to the vertical, or equivalently by μ = cos(θ). The atmosphere is assumed to be homogeneous and isothermal. The thermalization parameter entering Eq. (18) is chosen as ϵ = 10−4, while the damping constant entering the absorption profile ϕ as a = 10−3. A two-level atom with Ju = 1 and J = 0 is considered and the depolarizing effect of elastic collisions is neglected. All the problem parameters are summarized below

J u = 1 , J l = 0 , w 10 ( 0 ) = w 10 ( 2 ) = 1 , δ u ( K ) = 0 , a 10 ( 0 , 0 ) = a 10 ( 2 , 0 ) = 1 ϵ , W = 1 , ϵ = 10 4 , a = 10 3 , T 0 , 1 0 ( μ ) = 1 , T 0 , 2 0 ( μ ) = 0 , T 0 , 1 2 ( μ ) = 2 ( 3 μ 2 1 ) / 4 , T 0 , 2 2 ( μ ) = 2 ( 3 μ 2 3 ) / 4 . $$ \begin{aligned}&J_u = 1,\quad J_l = 0, \quad { w}^{(0)}_{10} = { w}^{(2)}_{10} = 1, \\&\delta _u^{(K)}=0, \quad a_{10}^{(0,0)} = a_{10}^{(2,0)} = 1 - \epsilon , \\&W = 1, \quad \epsilon =10^{-4}, \quad a=10^{-3}, \\&\mathcal{T} ^0_{0,1}(\mu )=1, \quad \mathcal{T} ^0_{0,2}(\mu )=0,\\&\mathcal{T} ^2_{0,1}(\mu )=\sqrt{2}(3\mu ^2-1)/4, \\&\mathcal{T} ^2_{0,2}(\mu )=\sqrt{2}(3\mu ^2-3)/4. \end{aligned} $$

In 1D geometries, the spatial vector r can be replaced by the scalar height coordinate z. Unless otherwise stated, hereafter, the atmosphere is discretized along the vertical direction through a grid in frequency-integrated optical depth (τ). In particular, a logarithmically spaced grid given by

10 5 = τ 1 < τ 2 < < τ N s = 10 4 $$ \begin{aligned} 10^{-5} = \tau _1 < \tau _2 < \cdots < \tau _{N_s} = 10^4 \end{aligned} $$

is considered in the following sections. The spectral line is sampled with Nν frequency nodes equally spaced in the reduced frequency interval [ − 5, 5], and NΩ Gauss-Legendre nodes are used to discretize μ ∈ [ − 1, 1].

For the sake of clarity, a limited number of formal solvers are analyzed: the first-order accurate implicit Euler method, the second-order accurate DELO-linear and DELOPAR methods, and the third-order accurate DELO-parabolic method5. However, the numerical analysis performed here can be unconditionally applied to any formal solver. Further details on the specific properties of the aforementioned formal solvers are given by Janett et al. (2017a,b). Similarly to Trujillo Bueno & Manso Sainz (1999), the block-Jacobi preconditioner is used with a block size of two that corresponds to the σ 0 0 $ \sigma^0_0 $ and σ 0 2 $ \sigma^2_0 $ components local coupling.

In the following sections, we take advantage of the algebraic formulation of the transfer problem (25) to study the convergence of iterative methods presented in Sect. 2. In particular, we focus on the spectral radius of the iteration matrix ρ(G), with G defined in Eq. (26), for various preconditioners P and formal solvers. We remark that the line thermal contributions (encoded in c) and boundary conditions (encoded in t) enter only in the right-hand side of Eq. (25) and, consequently, they do not affect the results of this study. Figure 1 displays the magnitude of the entries of the matrix IdJΛT for various formal solvers. Informally speaking, given the solution vector (cf. Appendix A)

σ = [ σ 0 0 ( τ 1 ) , σ 0 2 ( τ 1 ) , σ 0 0 ( τ 2 ) , σ 0 2 ( τ 2 ) , . . . , σ 0 0 ( τ N s ) σ 0 2 ( τ N s ) ] T , $$ \begin{aligned} \boldsymbol{\sigma } = [\sigma ^0_0(\tau _1), \sigma ^2_0(\tau _1),\sigma ^0_0(\tau _2), \sigma ^2_0(\tau _2),...,\sigma ^0_0(\tau _{N_s}) \sigma ^2_0(\tau _{N_s})]^T, \end{aligned} $$

thumbnail Fig. 1.

Values of log10|Id − JΛT|ij for i, j = 1, …, 2Ns and multiple formal solvers, with Ns = 80, NΩ = 21, and Nν = 20.

a large value of the entry |Id − JΛT|ij, with i, j ∈ {1, ..., 2Ns}, indicates that σi depends strongly on σj. The asymmetry of the matrices is mainly due to the usage of a nonuniform spatial grid. The matrices entries show that, as expected, for τ ≤ 1 (i.e., i < 90) the solution σ depends mostly on a region in the neighborhood of τ = 1. As τ increases this effect becomes less evident and the solution depends predominantly on neighboring nodes.

4.1. Impact of spectral and angular quadratures

Figure 2 illustrates that, once a minimal resolution is guaranteed, varying the number of nodes in the spectral and angular grids Nν and NΩ, respectively, has a negligible effect on ρ(G). Therefore, the setting Nν = 21 and NΩ = 20 remains fixed in the following numerical analysis.

thumbnail Fig. 2.

Spectral radius of the iteration matrix ρ(G) for the DELO-linear formal solver and the Richardson iterative method, with Ns = 40, varying NΩ and Nν. The same behavior is observed for the other iterative methods and formal solvers analyzed, as well as for different values of Ns.

4.2. Impact of the formal solver

We now present the convergence rates of various undamped (i.e., with ω = 1) iterative solvers, as a function of the number of spatial nodes for various formal solvers. The numerical results are summarized in Fig. 3. For the Richardson method (a.k.a. Λ−iteration) we observe that ρ(G)≈1, resulting in a very slow convergence. Indeed, this method is seldom used in practical applications. Moreover, we generally observe a lower convergence rates as Ns increases for all formal solvers. Among the considered formal solvers, DELOPAR shows the best convergence when combined with the Jacobi and Gauss-Seidel iterative methods. For the same iterative methods, DELOPAR is not convergent for Ns = 10, because in this case the matrix P−1(IdJΛT) has one negative eigenvalue, and thus Eq. (10) is not satisfied. A particular case is given by the DELO-parabolic, for which the Jacobi and block-Jacobi iterative methods never converge.

thumbnail Fig. 3.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and undamped iterative methods with NΩ = 20, Nν = 21, and varying Ns.

4.3. Impact of the damping parameter

We now discuss the impact of the damping parameter ω introduced in Sect. 2.3. Figure 4 reports ρ(G) for the damped methods as a function of the damping parameter ω and for multiple formal solvers. The top left panel shows that the impact of the damping parameter on the Richardson method is barely perceivable. Similarly, top right and bottom left panels indicates that the impact of the damping parameter on the Jacobi and block-Jacobi methods is not particularly pronounced. In fact, the spectral radius provided by the optimal parameter does not differ substantially from the value obtained with ω = 1. However, a quite interesting result is revealed: the Jacobi and block-Jacobi methods (at ω = 1) combined with the DELO-parabolic formal solver yield a spectral radius bigger than unity, preventing convergence. However, the use of a suitable damping parameter (ω < 0.81) enforces stability, guaranteeing convergence. An illustrative application of the damped block-Jacobi iterative method to transfer of polarized radiation is given by Janett et al. (2021). The bottom right panel confirms that the impact of the damping parameter in the SOR method is particularly effective. Indeed, a suitable choice of the damping parameter significantly reduces the spectral radius of the iteration matrix, increasing the convergence rate of the SOR method. This is consistent with the results of the previous investigations by Trujillo Bueno & Fabiani Bendicho (1995) and Trujillo Bueno & Manso Sainz (1999). We notice that the optimal damping parameter cannot be calculated with Eq. (12) for the DELO-parabolic case, since κ = ρ(GJac) is larger then one. Moreover, using the implicit Euler and DELOPAR formal solvers, the matrix GJac has few complex eigenvalues and therefore the estimate Eq. (12) is not sharp.

thumbnail Fig. 4.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and damped iterative methods with NΩ = 20, Nν = 21, Ns = 80, and varying the damping parameter ω. Vertical dashed lines represent the optimal damping parameters defined in Sect. 2.3.

Figure 5 collects the convergence rates of the damped Jacobi and SOR methods, using the optimal damping parameters ωopt and ωsor defined in Sect. 2.3.

thumbnail Fig. 5.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and for the damped-Jacobi and SOR iterative methods with NΩ = 20, Nν = 21, and varying Ns. For the damped-Jacobi and SOR methods, we used ωopt from Eq. (11) and ωsor from Eq. (12), respectively. Since Eq. (12) cannot be applied if κ < 1, the ωsor corresponding to DELOPAR is used for the DELO-parabolic formal solver.

4.4. Impact of the collisional destruction probability

Figure 6 reports convergence rates for multiple ϵ ∈ [10−6, 1), showing that a larger ϵ corresponds to faster convergence for all the iterative methods and formal solvers under investigation. This result indicates that the numerical solution becomes easier when collision effects prevail.

thumbnail Fig. 6.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and undamped iterative methods with NΩ = 20, Nν = 21, Ns = 40, and varying the collisional destruction probability ϵ.

5. Conclusions

This paper presents a fully algebraic formulation of a linear transfer problem for polarized radiation allowing for a formal analysis that aims to better understand the convergence properties of stationary iterative methods. In particular, we investigate how the convergence conditions of these methods depend on different problem and discretization settings, identifying cases where instability issues appear and exposing how to deal with them.

Numerical experiments confirm that the number of nodes in the spectral and angular grids has a negligible effect on the spectral radius of iteration matrices of various methods and, hence, on their convergence properties. By contrast, a larger number of spatial nodes leads to a reduction of convergence rates for all iterative methods and formal solvers under investigation.

In general, the use of damping parameters can both enforce stability and increase the convergence rate. Additionally, the optimal parameters described in Sect. 2.3 are indeed effective, especially in the SOR case. In the particular case, where the Jacobi and block-Jacobi methods are combined with the DELO-parabolic formal solver, the spectral radius of the iteration matrix becomes larger than unity. Hence, unless a suitable damping parameter is used, this combination is not convergent.

We remark that in this paper, we considered a specific benchmark linear radiative transfer problem. However, the presented methodology can also be applied to a wider range of settings, such as the unpolarized case, two-term atomic models including continuum contributions, 3D atmospheric models with arbitrary magnetic and bulk velocity fields, and theoretical frameworks accounting for PRD effects.

Moreover, the algebraic formulation of the transfer problem is used here exclusively as a practical tool for the convergence analysis. In the setting we have considered, the whole transfer problem is encoded in a dense linear system of size 2Ns. Different choices of the unknown vector (I, σ, or S) result in systems with different size, algebraic properties, and sparsity patterns.

However, this algebraic formulation paves the way to the application of advanced solution methods for linear (or nonlinear) systems, arising from radiative transfer applications. In particular, many state-of-the-art parallel numerical techniques can be employed, such as Krylov methods (see, e.g., Lambert et al. 2015), parallel preconditioners and multigrid acceleration (Hackbusch 2016; Saad 2003). Stationary iterative methods can also be applied as smoothers for multigrid techniques. These methods can also be applied in parallel, using available numerical libraries, which are tailored for high performance computing. In particular, Krylov methods are the subject of study of the second part of this paper series (Benedusi et al. 2021).


1

In the context of linear radiative transfer problems, the Richardson method corresponds to the well-known Λ-iteration, while Jacobi and block-Jacobi methods correspond to the so-called accelerated Λ-iteration (ALI) methods.

2

This is strictly true when J = 0. However, it is a good approximation when the lower level has a long lifetime (e.g., in case it is a ground or metastable level) and the rate of elastic depolarizing collisions is high (i.e., when the plasma density is high).

3

The statistical equilibrium equations additionally include

2 J + 1 ρ 0 , 0 ( r ) + 2 J u + 1 ρ 0 , u 0 ( r ) = 1 , $$ \begin{aligned} \sqrt{2J_\ell + 1}\rho ^0_{0,\ell }(\boldsymbol{r})+\sqrt{2J_u+1}\rho ^0_{0,u}(\boldsymbol{r})=1, \end{aligned} $$

which represents the conservation of the total number of atoms.

4

This quantity is given by

w J u J ( K ) = ( 1 ) 1 + J u + J 3 ( 2 J u + 1 ) { 1 1 K J u J u J } , $$ \begin{aligned} { w}^{(K)}_{J_u J_\ell } = (-1)^{1+J_u+J_\ell }\sqrt{3(2J_u+1)} \left\{ \begin{array}{c c c} 1&1&K \\ J_u&J_u&J_\ell \end{array} \right\} , \end{aligned} $$

where the quantity in curly parentheses is the 6j-symbol (see Landi Degl’Innocenti & Landolfi 2004, Chapt. 2).

5

We remark that the DELOPAR method effectively provides third-order accuracy in the case of a diagonal propagation matrix.

Acknowledgments

Special thanks are extended to E. Alsina Ballester, N. Guerreiro, and T. Simpson for particularly enriching comments on this work. The authors gratefully acknowledge the Swiss National Science Foundation (SNSF) for financial support through grant CRSII5_180238. Rolf Krause acknowledges the funding from the European High-Performance Computing Joint Undertaking (JU) under grant agreement No 955701 (project TIME-X). The JU receives support from the European Union’s Horizon 2020 research and innovation programme and Belgium, France, Germany, Switzerland.

References

  1. Alsina Ballester, E., Belluzzi, L., & Trujillo Bueno, J. 2017, ApJ, 836, 6 [Google Scholar]
  2. Alsina Ballester, E., Belluzzi, L., & Trujillo Bueno, J. 2018, ApJ, 854, 150 [Google Scholar]
  3. Belluzzi, L., & Trujillo Bueno, J. 2014, A&A, 564, A16 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  4. Benedusi, P., Janett, G., Belluzzi, L., & Krause, R. 2021, A&A, 655, A88 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  5. del Pino Alemán, T., Trujillo Bueno, J., Casini, R., & Manso Sainz, R. 2020, ApJ, 891, 91 [CrossRef] [Google Scholar]
  6. Hackbusch, W. 2016, in Iterative Solution of Large Sparse Systems of Equations, 2nd edn. (Springer), Appl. Math. Sci., 95 [CrossRef] [Google Scholar]
  7. Hageman, L. A., & Young, D. M. 1981, Applied Iterative Methods (New York, NY, USA: Academic) [Google Scholar]
  8. Hubeny, I. 2003, in Stellar Atmosphere Modeling, 288, 17 [NASA ADS] [Google Scholar]
  9. Janett, G., & Paganini, A. 2018, ApJ, 857, 91 [Google Scholar]
  10. Janett, G., Carlin, E. S., Steiner, O., & Belluzzi, L. 2017a, ApJ, 840, 107 [NASA ADS] [CrossRef] [Google Scholar]
  11. Janett, G., Steiner, O., & Belluzzi, L. 2017b, ApJ, 845, 104 [Google Scholar]
  12. Janett, G., Ballester, E. A., Guerreiro, N., et al. 2021, A&A, 655, A13 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  13. Klein, R. I., Castor, J. I., Greenbaum, A., Taylor, D., & Dykema, P. G. 1989, J. Quant. Spectr. Rad. Transf., 41, 199 [NASA ADS] [CrossRef] [Google Scholar]
  14. Lambert, J., Josselin, E., Ryde, N., & Faure, A. 2015, A&A, 580, A50 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  15. Lambert, J., Paletou, F., Josselin, E., & Glorian, J.-M. 2016, Eur. J. Phys., 37, 015603 [Google Scholar]
  16. Landi Degl’Innocenti, E., & Landolfi, M. 2004, in Polarization in Spectral Lines, (Dordrecht: Kluwer Academic Publishers), Astrophys. Space Sci. Lib., 307 [Google Scholar]
  17. Olson, G. L., Auer, L. H., & Buchler, J. R. 1986, J. Quant. Spectr. Rad. Transf., 35, 431 [Google Scholar]
  18. Paletou, F., & Auer, L. H. 1995, A&A, 297, 771 [NASA ADS] [Google Scholar]
  19. Quarteroni, A., Sacco, R., & Saleri, F. 2010, Numerical Mathematics (Springer Science& Business Media), 37 [Google Scholar]
  20. Rybicki, G. B., & Hummer, D. G. 1991, A&A, 245, 171 [NASA ADS] [Google Scholar]
  21. Saad, Y. 2003, Iterative Methods for Sparse Linear Systems (SIAM) [Google Scholar]
  22. Sampoorna, M., & Trujillo Bueno, J. 2010, ApJ, 712, 1331 [NASA ADS] [CrossRef] [Google Scholar]
  23. Sampoorna, M., Nagendra, K. N., & Stenflo, J. O. 2017, ApJ, 844, 97 [NASA ADS] [CrossRef] [Google Scholar]
  24. Trujillo Bueno, J., & Fabiani Bendicho, P. 1995, ApJ, 455, 646 [NASA ADS] [CrossRef] [Google Scholar]
  25. Trujillo Bueno, J., & Landi Degl’Innocenti, E. 1996, Sol. Phys., 164, 135 [NASA ADS] [CrossRef] [Google Scholar]
  26. Trujillo Bueno, J., & Manso Sainz, R. 1999, ApJ, 516, 436 [NASA ADS] [CrossRef] [Google Scholar]

Appendix A: Matrix J

The vector I ∈ ℝ4NsNΩNν collects all the discrete values of the Stokes parameters with the order defined by the tensor notation I ∈ ℝNs × ℝ4 × ℝNΩ × ℝNν, namely,

I = [ I 1 ( r 1 , Ω 1 , ν 1 ) , I 1 ( r 1 , Ω 1 , ν 2 ) , , I 1 ( r 1 , Ω 1 , ν N ν ) , I 1 ( r 1 , Ω 2 , ν 1 ) , , I 1 ( r 1 , Ω 2 , ν N ν ) , , I 1 ( r 1 , Ω N Ω , ν N ν ) , I 2 ( r 1 , Ω 1 , ν 1 ) , , I 2 ( r 1 , Ω N Ω , ν N ν ) , , I 4 ( r 1 , Ω N Ω , ν N ν ) , I 1 ( r 2 , Ω 1 , ν 1 ) , , I 1 ( r 2 , Ω N Ω , ν N ν ) , , I 4 ( r N s , Ω N Ω , ν N ν ] T . $$ \begin{aligned} \mathbf I =&[ I_1(\mathbf{r }_1,\mathbf {\Omega} _1,\nu _1), I_1(\mathbf{r }_1,\mathbf {\Omega} _1,\nu _2),\ldots , I_1(\mathbf{r }_1,\mathbf {\Omega} _1,\nu _{N_\nu }), I_1(\mathbf{r}_1,\mathbf {\Omega} _2,\nu _1),\ldots , I_1(\mathbf{r }_1,\mathbf {\Omega} _2,\nu _{N_\nu }),\ldots , I_1(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu }), I_2(\mathbf{r }_1,\mathbf {\Omega} _1,\nu _1),\nonumber \\&\qquad \ldots , I_2(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu }),\ldots , I_4(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu }), I_1(\mathbf{r }_2,\mathbf {\Omega} _1,\nu _1),\ldots , I_1(\mathbf{r }_2,\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu }),\ldots , I_4(\mathbf{r }_{N_s},\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu }]^T. \end{aligned} $$(A.1)

We also define the vector I k R 4 N Ω N ν $ \mathbf{I}_k\in\mathbb R^{4N_\Omega N_\nu} $ given by

I k = [ I 1 ( r k , Ω 1 , ν 1 ) , , I 4 ( r k , Ω N Ω , ν N ν ) ] T , $$ \begin{aligned} \mathbf I _k=\left[I_1(\mathbf{r }_k,\mathbf {\Omega} _1,\nu _1),\ldots ,I_4(\mathbf{r }_k,\mathbf {\Omega} _{N_\Omega },\nu _{N_\nu })\right]^T, \end{aligned} $$

which collects the entries of I corresponding to the spatial point rk, such that I = [ I 1 T ,, I N s T ] T $ \mathbf{I}=\left[\mathbf{I}_1^T,\ldots,\mathbf{I}_{N_s}^T\right]^T $.

For K = 0, 1, 2, Q = −K, …, K and k = 1, …, Ns we express (18) and (19) as

σ Q K ( r k ) = i = 1 4 m = 1 N Ω p = 1 N ν J K Q , i ( r k , Ω m , ν p ) I i ( r k , Ω m , ν p ) + c Q K ( r k ) = a J u J ( K , Q ) ( r k ) k Q K I k + c Q K ( r k ) , $$ \begin{aligned} \sigma ^K_Q(\mathbf{r }_k) = \sum _{i=1}^4\sum _{m=1}^{N_\Omega }\sum _{p=1}^{N_\nu } J_{KQ,i}(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p)I_i(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p)+ c_Q^K(\mathbf{r }_k)= a_{J_u J_\ell }^{(K,Q)}(\mathbf{r }_k)\mathbf k _{-Q}^K\mathbf I _k + c_Q^K(\mathbf{r }_k), \end{aligned} $$(A.2)

where

J K Q , i ( r k , Ω m , ν p ) = a J u J ( K , Q ) ( r k ) k Q , N ν N Ω ( i 1 ) + N ν ( m 1 ) + p K . $$ \begin{aligned} J_{KQ,i}(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p) = a_{J_u J_\ell }^{(K,Q)}(\mathbf{r }_k) \mathbf k ^K_{-Q,N_\nu N_\Omega (i-1) + N_\nu (m-1) + p}. \end{aligned} $$(A.3)

The row vector k Q K R 4 N Ω N ν $ \mathbf{k}_{-Q}^K\in\mathbb R^{4N_\Omega N_\nu} $ collects the quadrature coefficients for the integral in (14) and is given by the Kronecker product

k Q K = 1 4 π ψ Q K ϕ , $$ \begin{aligned} \mathbf k _Q^K = \frac{1}{4\pi }\boldsymbol{\psi }_Q^K\otimes \boldsymbol{\phi }, \end{aligned} $$(A.4)

where ψ Q K 4 N Ω $ \boldsymbol{\psi}^K_Q\in\mathbb{R}^{4N_\Omega} $ and ϕ ∈ ℝNν are given by

ψ Q , ( i 1 ) N Ω + m K = T Q , i K ( Ω m ) u m , for i = 1 , , 4 and m = 1 , , N Ω , ϕ p = ϕ ( ν p ) v p , for p = 1 , , N ν , $$ \begin{aligned} \boldsymbol{\psi }^K_{Q,(i-1)N_\Omega +m}&= \mathcal{T} ^K_{Q,i}(\mathbf{\Omega }_m) u_m, \text{ for } i=1,\ldots ,4 \text{ and } m = 1,\ldots ,N_\Omega ,\\ \boldsymbol{\phi }_p&= \phi (\nu _p)v_p, \text{ for } p = 1,\ldots ,N_\nu , \end{aligned} $$

and um and vp are the quadrature weights of the grids { Ω m } m=1 N Ω $ \{\mathbf{\Omega}_m\}_{m=1}^{N_\Omega} $ and { ν p } p=1 N ν $ \{\nu_p\}_{p=1}^{N_{\nu}} $, respectively.

The matrix J R 9 N s × 4 N s N Ω N ν $ J\in\mathbb R^{9 N_s\times4 N_s N_\Omega N_\nu} $ appearing in (22) is then given by

J = [ J ¯ ( r 1 ) J ¯ ( r 2 ) J ¯ ( r N s ) ] , where J ¯ ( r k ) = [ a J u J ( 0 , 0 ) ( r k ) k 0 0 a J u J ( 1 , 1 ) ( r k ) k 1 1 a J u J ( 1 , 0 ) ( r k ) k 0 1 a J u J ( 2 , 1 ) ( r k ) k 1 2 a J u J ( 2 , 2 ) ( r k ) k 2 2 ] . $$ J = \begin{bmatrix} \bar{J}(\mathbf{r }_1)&&\\&\bar{J}(\mathbf{r }_2)&\\&\ddots&\\&&\bar{J}(\mathbf{r }_{N_s}) \\ \end{bmatrix}, \qquad \text{ where} \qquad \bar{J}(\mathbf{r }_k) = \begin{bmatrix} a_{J_u J_\ell }^{(0,0)}(\mathbf{r }_k)\mathbf k _0^0 \\ a_{J_u J_\ell }^{(1,-1)}(\mathbf{r }_k)\mathbf k ^1_1\\ a_{J_u J_\ell }^{(1,0)}(\mathbf{r }_k)\mathbf k ^1_0\\ \vdots \\ a_{J_u J_\ell }^{(2,1)}(\mathbf{r }_k)\mathbf k ^2_{-1} \\ a_{J_u J_\ell }^{(2,2)}(\mathbf{r }_k)\mathbf k ^2_{-2} \\ \end{bmatrix}. $$

The matrix J ¯ ( r k ) R 9 × 4 N Ω N ν $ \bar{J}({\mathbf{r}}_k)\in\mathbb R^{9\times4N_\Omega N_\nu} $, which collects the coefficients a J u J ( K , Q ) ( r k ) k Q K $ a_{J_u J_\ell}^{(K,Q)}({\mathbf{r}}_k)\mathbf{k}_{-Q}^K $ is generally dense, but it also contains zero blocks, because T Q,i K ( Ω m ) $ \mathcal{T}^K_{Q,i}({\bf\Omega}_m) $ is null for certain triples i, K, Q.

Accordingly to the structure of J, the vectors σ , c R 9 N s $ {\boldsymbol{\sigma}},\mathbf{c}\in\mathbb R^{9N_s} $ appearing in (22) are given by

c = [ c ( r 1 ) , c ( r 2 ) , , c ( r N s ) ] T with c ( r k ) = [ c 0 0 ( r k ) , c 1 1 ( r k ) , c 0 1 ( r k ) , , c 2 2 ( r k ) ] R 9 , σ = [ σ ( r 1 ) , σ ( r 2 ) , , σ ( r N s ) ] T with σ ( r k ) = [ σ 0 0 ( r k ) , σ 1 1 ( r k ) , σ 0 1 ( r k ) , , σ 2 2 ( r k ) ] R 9 . $$ \begin{aligned}&\mathbf c = \left[\tilde{\mathbf{c }}(\mathbf{r }_1),\tilde{\mathbf{c }}(\mathbf{r }_2),\ldots ,\tilde{\mathbf{c }}(\mathbf{r }_{N_s})\right]^T \quad \text{ with} \quad \tilde{\mathbf{c }}(\mathbf{r }_k) = \left[c_0^0(\mathbf{r }_k),c_{-1}^1(\mathbf{r }_k),c_0^1(\mathbf{r }_k),\ldots ,c_2^2(\mathbf{r }_k)\right]\in \mathbb{ R} ^9,\\&\boldsymbol{\sigma } = \left[\tilde{\boldsymbol{\sigma }}(\mathbf{r }_1),\tilde{\boldsymbol{\sigma }}(\mathbf{r }_2),\ldots ,\tilde{\boldsymbol{\sigma }}(\mathbf{r }_{N_s})\right]^T \quad \text{ with} \quad \tilde{\boldsymbol{\sigma }}(\mathbf{r }_k) = \left[\sigma ^0_0(\mathbf{r }_k),\sigma _{-1}^1(\mathbf{r }_k),\sigma _0^1(\mathbf{r }_k),\ldots ,\sigma _2^2(\mathbf{r }_k)\right]\in \mathbb{ R} ^9. \end{aligned} $$

For completeness, we express the entries of the matrix J in term of the coefficient appearing in (19), namely,

J K Q , i ( r k , Ω m , ν p ) = J l l , with l = 9 ( k 1 ) + K ( K + 1 ) + Q + 1 and l = 4 N Ω N ν ( k 1 ) + N Ω N ν ( i 1 ) + N ν ( m 1 ) + p . $$ J_{KQ,i}(\mathbf{r }_k,\mathbf{\Omega }_m,\nu _p) = J_{ll^{\prime }}, \quad \text{ with} \quad l=9(k-1) + K(K+1) + Q + 1 \quad \text{ and} \quad l^{\prime }=4N_\Omega N_\nu (k-1) + N_\Omega N_\nu (i-1) + N_\nu (m-1) + p. $$

Appendix B: Matrix T

The vector S ∈ ℝ4NsNΩ collects all the discrete values of the source function vector with the order defined by the tensor notation S ∈ ℝNs × ℝ4 × ℝNΩ, namely,

S = [ S 1 ( r 1 , Ω 1 ) , S 1 ( r 1 , Ω 2 ) , , S 1 ( r 1 , Ω N Ω ) , S 2 ( r 1 , Ω 1 ) , , S 2 ( r 1 , Ω N Ω ) , , S 4 ( r 1 , Ω N Ω ) , S 1 ( r 2 , Ω 1 ) , , S 4 ( r N s , Ω N Ω ) ] T . $$ \begin{aligned} \mathbf S = [S_1(\mathbf{r }_1,\mathbf {\Omega} _1), S_1(\mathbf{r }_1,\mathbf {\Omega} _2),\ldots , S_1(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega }), S_2(\mathbf{r }_1,\mathbf {\Omega} _1),\ldots , S_2(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega }),\ldots , S_4(\mathbf{r }_1,\mathbf {\Omega} _{N_\Omega }), S_1(\mathbf{r }_2,\mathbf {\Omega} _1),\ldots , S_4(\mathbf{r }_{N_s},\mathbf {\Omega} _{N_\Omega }) ]^T. \end{aligned} $$

We also define the auxiliary vector w i ( Ω ) R 9 $ \mathbf{w}_i(\mathbf{\Omega})\in\mathbb R^9 $ given by

w i ( Ω ) = [ w J u J ( 0 ) T 0 0 ( i , Ω ) , w J u J ( 1 ) T 1 1 ( i , Ω ) , w J u J ( 1 ) T 0 1 ( i , Ω ) , , w J u J ( 2 ) T 1 2 ( i , Ω ) , w J u J ( 2 ) T 2 2 ( i , Ω ) ] . $$ \begin{aligned} \mathbf w _i(\mathbf {\Omega} ) = \left[ w^{(0)}_{J_u J_\ell } \mathcal{T} ^0_0(i,\mathbf{\Omega }), w^{(1)}_{J_u J_\ell }\mathcal{T} ^1_{-1}(i,\mathbf{\Omega }), w^{(1)}_{J_u J_\ell } \mathcal{T} ^1_0(i,\mathbf{\Omega }),\ldots , w^{(2)}_{J_u J_\ell } \mathcal{T} ^2_1(i,\mathbf{\Omega }), w^{(2)}_{J_u J_\ell } \mathcal{T} ^2_2(i,\mathbf{\Omega }) \right]. \end{aligned} $$

The matrix T R 4 N s N Ω × 9 N s $ T\in\mathbb R^{4 N_s N_\Omega \times9 N_s} $ appearing in (23) is then given by

T = [ W W W ] , where W = [ w 1 ( Ω 1 ) w 1 ( Ω 2 ) w 1 ( Ω N Ω ) w 2 ( Ω 1 ) w 4 ( Ω N Ω ) ] . $$ T = \begin{bmatrix} W&&\\&W&\\&\ddots&\\&&W \\ \end{bmatrix}, \qquad \text{ where} \qquad W = \begin{bmatrix} \mathbf w _1(\mathbf{\Omega }_1) \\ \mathbf w _1(\mathbf{\Omega }_2) \\ \vdots \\ \mathbf w _1(\mathbf{\Omega }_{N_\Omega }) \\ \mathbf w _2(\mathbf{\Omega }_1) \\ \vdots \\ \mathbf w _4(\mathbf{\Omega }_{N_\Omega }) \\ \end{bmatrix}. $$

The matrix W R 4 N Ω × 9 $ W\in\mathbb R^{4N_\Omega\times9} $ is generally dense and the matrix T is also given by the Kronecker product T = I d N s W $ T =Id_{N_s} \otimes W $, where I d N s $ Id_{N_s} $ is the identity matrix of size Ns.

Appendix C: Matrix Λ

For the sake of clarity, the implicit Euler method is used to integrate the transfer equation (15). However, the same analysis can be generalized to any formal solver. For simplicity, we consider the 1D plane-parallel cylindrical symmetric atmospheric model used in Section 4.

The angular dependence of the problem is encoded by the scalar μ = cos(θ) with the corresponding discrete grid

1 μ 1 < μ 2 < < μ N Ω 1 < μ N Ω 1 . $$ \begin{aligned} -1 \le \mu _1 < \mu _2 < \ldots < \mu _{N_\Omega -1} < \mu _{N_\Omega } \le 1. \end{aligned} $$

The spatial dependence of the problem is encoded either by the scalar s with the corresponding discrete grid { s k } k=1 N s $ \{s_k\}_{k=1}^{N_s} $ or by the scaled line-center frequency optical depth τ with the corresponding discrete grid { τ k } k=1 N s $ \{\tau_k\}_{k=1}^{N_s} $. The relation between the two spatial variables is described by

ϕ ( ν ) d τ = η ( s , ν ) d s $$ \phi (\nu )\mathrm{d} \tau =-\eta (s,\nu )\mathrm{d} s $$

and the transfer equation (15) can then be expressed as

μ ϕ ( ν ) d d τ I i ( τ , μ , ν ) = I i ( τ , μ , ν ) S i ( τ , μ ) . $$ \begin{aligned} \frac{\mu }{\phi (\nu )}\frac{\mathrm{d} }{\mathrm{d} \tau } I_i(\tau ,\mu ,\nu ) = I_i(\tau ,\mu ,\nu ) - S_{\!i}(\tau ,\mu ). \end{aligned} $$(C.1)

Applying the implicit Euler method to (C.1) for i = 1, …, 4, we obtain, for incoming and outgoing directions respectively

I i ( τ k + 1 , μ , ν ) = I i ( τ k , μ , ν ) + Δ τ k ( μ , ν ) S i ( τ k + 1 , μ ) 1 + Δ τ k ( μ , ν ) , for μ < 0 and k = 1 , , N s 1 , $$ \begin{aligned} I_i(\tau _{k+1},\mu ,\nu )&= \frac{I_i(\tau _{k},\mu ,\nu ) +\Delta \tau _k(\mu ,\nu )S_{\!i}(\tau _{k+1},\mu )}{1+\Delta \tau _k(\mu ,\nu )},\qquad \text{ for} \quad \mu < 0 \quad \text{ and} \quad k = 1,\ldots ,N_{s}-1,\end{aligned} $$(C.2)

I i ( τ k 1 , μ , ν ) = I i ( τ k , μ , ν ) + Δ τ k 1 ( μ , ν ) S i ( τ k 1 , μ ) 1 + Δ τ k 1 ( μ , ν ) , for μ > 0 and k = N s , N s 1 , , 2 , $$ \begin{aligned} I_i(\tau _{k-1},\mu ,\nu )&= \frac{I_i(\tau _k,\mu ,\nu ) +\Delta \tau _{k-1}(\mu ,\nu )S_{\!i}(\tau _{k-1},\mu )}{1+\Delta \tau _{k-1}(\mu ,\nu )}, \qquad \text{ for} \quad \mu >0 \quad \text{ and} \quad k = N_s,N_{s}-1,\ldots ,2, \end{aligned} $$(C.3)

where

Δ τ k ( μ , ν ) = ϕ ( ν ) | μ | | τ k + 1 τ k | . $$ \begin{aligned} \Delta \tau _{k}(\mu ,\nu )=\frac{\phi (\nu )}{|\mu |}|\tau _{k+1}-\tau _k|. \end{aligned} $$

We notice that different Stokes parameters, directions and frequencies are decoupled in the transfer equation (15) and, consequently, in the discrete counterparts (C.2) and (C.3). Providing the initial conditions I i ( τ 1 ,μ,ν)= I i in (μ,ν) $ I_i(\tau_1,\mu,\nu)=I_i^{\text{in}}(\mu,\nu) $ for μ < 0 and I i ( τ N s ,μ,ν)= I i out (μ,ν) $ I_i(\tau_{N_s},\mu,\nu)=I_i^{\text{out}}(\mu,\nu) $ for μ > 0, we recursively apply (C.2) and (C.3) and obtain

I i ( τ k , μ , ν ) = f k ( μ , ν ) I i in ( μ , ν ) + j = 2 k f k , j ( μ , ν ) S i ( τ j , μ ) for μ < 0 and k = 2 , , N s , $$ \begin{aligned} I_i(\tau _k,\mu ,\nu )&= f_k(\mu ,\nu ) I_i^{\text{ in}}(\mu ,\nu ) + \sum _{j=2}^k f_{k,j}(\mu ,\nu )S_{\!i}(\tau _j,\mu ) \qquad \text{ for} \quad \mu < 0 \quad \text{ and} \quad k = 2,\ldots ,N_s,\end{aligned} $$(C.4)

I i ( τ k , μ , ν ) = h k ( μ , ν ) I i out ( μ , ν ) + j = k N s 1 h k , j ( μ , ν ) S i ( τ j , μ ) for μ > 0 and k = 1 , , N s 1 , $$ \begin{aligned} I_i(\tau _k,\mu ,\nu )&= h_k(\mu ,\nu ) I_i^{\text{ out}}(\mu ,\nu ) + \sum _{j=k}^{N_s-1} h_{k,j}(\mu ,\nu )S_{\!i}(\tau _j,\mu ) \qquad \text{ for} \quad \mu >0 \quad \text{ and} \quad k = 1,\ldots ,N_s-1, \end{aligned} $$(C.5)

with

f k , j ( μ , ν ) = Δ τ j 1 ( μ , ν ) l = j k 1 + Δ τ l 1 ( μ , ν ) and f k ( μ , ν ) = 1 l = 1 k 1 1 + Δ τ l ( μ , ν ) , $$ \begin{aligned} f_{k,j}(\mu ,\nu )&= \frac{\Delta \tau _{j-1}(\mu ,\nu )}{\prod _{l=j}^{k} 1 + \Delta \tau _{l-1}(\mu ,\nu )} \quad \text{ and} \quad f_k(\mu ,\nu ) = \frac{1}{\prod _{l=1}^{k-1} 1 + \Delta \tau _{l}(\mu ,\nu )},\end{aligned} $$(C.6)

h k , j ( μ , ν ) = Δ τ j ( μ , ν ) l = k j 1 + Δ τ l ( μ , ν ) and h k ( μ , ν ) = 1 l = k N s 1 1 + Δ τ l ( μ , ν ) . $$ \begin{aligned} h_{k,j}(\mu ,\nu )&= \frac{\Delta \tau _{j}(\mu ,\nu )}{\prod _{l=k}^{j} 1 + \Delta \tau _{l}(\mu ,\nu )} \quad \text{ and} \quad h_k(\mu ,\nu ) = \frac{1}{\prod _{l=k}^{N_s-1} 1 + \Delta \tau _{l}(\mu ,\nu )}. \end{aligned} $$(C.7)

Equation (C.4) for i = 1, …, 4 can be arranged in a linear system with a lower triangular matrix F R N s × N s $ F\in\mathbb R^{N_s\times N_s} $, namely,

[ I i ( τ 1 , μ , ν ) I i ( τ 2 , μ , ν ) I i ( τ N s , μ , ν ) ] = [ 0 & f 2 , 2 ( μ , ν ) f N s , 2 ( μ , ν ) f N s , N s ( μ , ν ) ] [ S i ( τ 1 , μ ) S i ( τ 2 , μ ) S i ( τ N s , μ ) ] + [ 1 f 2 ( μ , ν ) f N s ( μ , ν ) ] I i in ( μ , ν ) I i ( μ , ν ) = F ( μ , ν ) S i ( μ ) + f i ( μ , ν ) . $$ \begin{bmatrix} I_i(\tau _1,\mu ,\nu ) \\ I_i(\tau _2,\mu ,\nu )\\ \vdots \\ I_i(\tau _{N_s},\mu ,\nu ) \\ \end{bmatrix} = \begin{bmatrix} 0&&\\\&f_{2,2}(\mu ,\nu )&\\&\vdots&\ddots&\\&f_{N_s,2}(\mu ,\nu )&\cdots&f_{N_s,N_s}(\mu ,\nu )\\ \end{bmatrix} \begin{bmatrix} S_{\!i}(\tau _1,\mu ) \\ S_{\!i}(\tau _2,\mu ) \\ \vdots \\ S_{\!i}(\tau _{N_s},\mu ) \\ \end{bmatrix} + \begin{bmatrix} 1 \\ f_2(\mu ,\nu ) \\ \vdots \\ f_{N_s}(\mu ,\nu ) \\ \end{bmatrix}I_i^{\text{ in}}(\mu ,\nu ) \Longleftrightarrow \tilde{\mathbf{I }}_i(\mu ,\nu ) = F(\mu ,\nu )\mathbf S _i(\mu ) + \mathbf{f }_i(\mu ,\nu ). $$

Analogously, Equation (C.5) for i = 1, …, 4 can be arranged in a linear system with an upper triangular matrix H R N s × N s $ H\in\mathbb R^{N_s\times N_s} $, namely,

[ I i ( τ 1 , μ , ν ) I i ( τ 2 , μ , ν ) I i ( τ N s , μ , ν ) ] = [ h 1 , 1 ( μ , ν ) h 1 , N s 1 ( μ , ν ) h N s 1 , N s 1 ( μ , ν ) 0 ] [ S i ( τ 1 , μ ) S i ( τ 2 , μ ) S i ( τ N s , μ ) ] + [ h 1 ( μ , ν ) h N s 1 ( μ , ν ) 1 ] I i out ( μ , ν ) I i ( μ , ν ) = H ( μ , ν ) S i ( μ ) + h i ( μ , ν ) . $$ \begin{bmatrix} I_i(\tau _1,\mu ,\nu ) \\ I_i(\tau _2,\mu ,\nu )\\ \vdots \\ I_i(\tau _{N_s},\mu ,\nu ) \\ \end{bmatrix} = \begin{bmatrix} h_{1,1}(\mu ,\nu )&\cdots&\cdots h_{1,N_s- 1}(\mu ,\nu )&\\&\ddots&\vdots&\\&h_{N_s-1,N_s-1}(\mu ,\nu )&\\&&0 \end{bmatrix} \begin{bmatrix} S_{\!i}(\tau _1,\mu ) \\ S_{\!i}(\tau _2,\mu ) \\ \vdots \\ S_{\!i}(\tau _{N_s},\mu ) \\ \end{bmatrix} + \begin{bmatrix} h_1(\mu ,\nu ) \\ \vdots \\ h_{N_s-1}(\mu ,\nu ) \\ 1 \\ \end{bmatrix}I_i^{\text{ out}}(\mu ,\nu ) \Longleftrightarrow \tilde{\mathbf{I }}_i(\mu ,\nu ) = H(\mu ,\nu )\mathbf S _i(\mu )+\mathbf h _i(\mu ,\nu ). $$

The two previous linear systems can then be combined into a single linear system of size Ns × Ns given by

I i ( μ , ν ) = L ( μ , ν ) S i ( μ ) + l i ( μ , ν ) , $$ \begin{aligned} \tilde{\mathbf{I }}_i(\mu ,\nu ) = L(\mu ,\nu )\mathbf S _i(\mu )+\mathbf l _i(\mu ,\nu ), \end{aligned} $$(C.8)

with

L ( μ , ν ) = { F ( μ , ν ) if μ < 0 , H ( μ , ν ) if μ > 0 , and l i ( μ , ν ) = { f i ( μ , ν ) if μ < 0 , h i ( μ , ν ) if μ > 0 . $$ \begin{aligned} L(\mu ,\nu ) = {\left\{ \begin{array}{ll} F(\mu ,\nu )&\text{ if } \mu < 0, \\ H(\mu ,\nu )&\text{ if } \mu > 0, \end{array}\right.} \qquad \text{ and} \qquad \mathbf l _i(\mu ,\nu ) = {\left\{ \begin{array}{ll} \mathbf f _i(\mu ,\nu )&\text{ if } \mu < 0, \\ \mathbf h _i(\mu ,\nu )&\text{ if } \mu > 0. \end{array}\right.} \end{aligned} $$

For k, l = 1, ..., Ns, m = 1, NΩ and p = 1, ..,Nν, the entries of the matrix Λ R 4 N s N Ω N ν × 4 N s N Ω $ \Lambda\in\mathbb R^{4 N_s N_\Omega N_\nu \times4 N_s N_\Omega } $ appearing in (24) and in (21), are then given by

Λ k l = Λ i ( τ k , τ l , μ m , ν p ) = ( L ( μ m , ν p ) ) k , l with k = 4 N Ω N ν ( k 1 ) + N Ω N ν ( i 1 ) + N ν ( m 1 ) + p , l = 4 N Ω ( l 1 ) + N Ω ( i 1 ) + m . $$ \begin{aligned}&\Lambda _{k^{\prime }l^{\prime }} =\Lambda _i(\tau _k,\tau _l,\mu _m,\nu _p)=\left(L(\mu _m,\nu _p)\right)_{k,l} \qquad \text{ with}&k^{\prime } = 4N_\Omega N_\nu (k-1) + N_\Omega N_\nu (i-1) + N_\nu (m-1) + p, \\&l^{\prime }= 4N_\Omega (l-1) + N_\Omega (i-1) + m.\nonumber \end{aligned} $$(C.9)

Accordingly, the vector t appearing in (24) reads

t = [ t 1 ( τ 1 , μ 1 , ν 1 ) , , t 4 ( τ 1 , μ N Ω , ν N ν ) , t 1 ( τ 2 , μ 1 , ν 1 ) , , t 4 ( τ 2 , μ N Ω , ν N ν ) , t 1 ( τ 3 , μ 1 , ν 1 ) , , t 4 ( τ N s , μ N Ω , ν N ν ) ] T , $$ \begin{aligned} \mathbf t = \left[ t_1(\tau _1,\mu _1,\nu _1),\ldots , t_4(\tau _1,\mu _{N_\Omega },\nu _{N_\nu }), t_1(\tau _2,\mu _1,\nu _1),\ldots , t_4(\tau _2,\mu _{N_\Omega },\nu _{N_\nu }), t_1(\tau _3,\mu _1,\nu _1),\ldots , t_4(\tau _{N_s},\mu _{N_\Omega },\nu _{N_\nu }) \right]^T, \end{aligned} $$

with

t i ( τ k , μ m , ν p ) = ( l i ( μ m , ν p ) ) k . $$ \begin{aligned} t_i(\tau _k,\mu _m,\nu _p) = \left(\mathbf l _i(\mu _m,\nu _p)\right)_k. \end{aligned} $$

that is using the same ordering of (A.1).

For the sake of clarity, we only considered the implicit Euler method for the explicit assembly of the matrix Λ. However, this analysis can be unconditionally applied to any formal solver, such as exponential integrators or multistep methods. In practice, it would be sufficient to modify (C.6) and (C.7) according to the selected formal solver.

All Figures

thumbnail Fig. 1.

Values of log10|Id − JΛT|ij for i, j = 1, …, 2Ns and multiple formal solvers, with Ns = 80, NΩ = 21, and Nν = 20.

In the text
thumbnail Fig. 2.

Spectral radius of the iteration matrix ρ(G) for the DELO-linear formal solver and the Richardson iterative method, with Ns = 40, varying NΩ and Nν. The same behavior is observed for the other iterative methods and formal solvers analyzed, as well as for different values of Ns.

In the text
thumbnail Fig. 3.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and undamped iterative methods with NΩ = 20, Nν = 21, and varying Ns.

In the text
thumbnail Fig. 4.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and damped iterative methods with NΩ = 20, Nν = 21, Ns = 80, and varying the damping parameter ω. Vertical dashed lines represent the optimal damping parameters defined in Sect. 2.3.

In the text
thumbnail Fig. 5.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and for the damped-Jacobi and SOR iterative methods with NΩ = 20, Nν = 21, and varying Ns. For the damped-Jacobi and SOR methods, we used ωopt from Eq. (11) and ωsor from Eq. (12), respectively. Since Eq. (12) cannot be applied if κ < 1, the ωsor corresponding to DELOPAR is used for the DELO-parabolic formal solver.

In the text
thumbnail Fig. 6.

Spectral radius of the iteration matrix ρ(G) for multiple formal solvers and undamped iterative methods with NΩ = 20, Nν = 21, Ns = 40, and varying the collisional destruction probability ϵ.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.