Optimal point spread function modeling for weak lensing: complexity and sparsity
S. PaulinHenriksson^{1}  A. Refregier^{1}  A. Amara^{2,3}
1  Service d'Astrophysique, CEA Saclay, Bâtiment 709, 91191 GifsurYvette Cedex, France
2 
Department of Physics, ETH Zürich, WolfgangPauliStrasse 16, 8093 Zürich,
Switzerland
3 
Department of Physics and Center for Theoretical and Computational Physics, University of Hong Kong, Pok Fu Lam Road, Hong Kong
Received 30 September 2008 / Accepted 23 December 2008
Abstract
Context. We address the issue of controling the systematic errors in shape measurements for weak gravitational lensing.
Aims. We make a step to quantify the impact of systematic errors in modeling the point spread function (PSF) of observations, on the determination of cosmological parameters from cosmic shear.
Methods. We explore the impact of PSF fitting errors on cosmic shear measurements using the concepts of complexity and sparsity. Complexity, introduced in a previous paper, characterizes the number of degrees of freedom of the PSF. For instance, fitting an underlying PSF with a model of low complexity produces small statistical errors on the model parameters, although these parameters could be affected by large biases. Alternatively, fitting a large number of parameters (i.e. a high complexity) tends to reduce biases at the expense of increasing the statistical errors. We attempt to find a tradeoff between scatters and biases by studying the mean squared error of a PSF model. We also characterize the model sparsity, which describes how efficiently the model is able to represent the underlying PSF using a limited number of free parameters. We present the general case and give an illustration for a realistic example of a PSF modeled by a shapelet basis set.
Results. We derive a relation between the complexity and the sparsity of the PSF model, the signaltonoise ratio of stars and the systematic errors in the cosmological parameters. By insisting that the systematic errors are below the statistical uncertainties, we derive a relation between the number of stars required to calibrate the PSF and the sparsity of the PSF model. We discuss the impact of our results for current and future cosmic shear surveys. In the typical case where the sparsity can be represented by a powerlaw function of the complexity, we demonstrate that current groundbased surveys can calibrate the PSF with few stars, while future surveys will require hard constraints on the sparsity in order to calibrate the PSF with 50 stars.
Key words: gravitational lensing  cosmology: dark matter  cosmology: cosmological parameters  cosmology: observations
1 Introduction
Studying spatial correlations between galaxy shapes induced by gravitational lensing of the large scale structure (``Cosmic Shear'') is a powerful probe of dark energy and dark matter. A number of current and planned surveys are dedicated to cosmic shear, such as: the CanadaFranceHawaiiTelescope Legacy Survey^{} (CFHTLS), the KIlo Degree Survey and the VISTA KiloDegree Infrared Galaxy Survey^{} (KIDS/VIKING), the Dark Energy Survey^{} (DES), the Panoramic Survey Telescope & Rapid Response System^{} (PanSTARRS), the SuperNovae Acceleration Probe^{} (SNAP), the Large Synoptic Survey Telescope^{} (LSST) and the Dark UNiverse Explorer^{} (DUNE/Euclid).
The most efficient way of improving the statistical precision of cosmic shear analyses is by enlarging the surveys. As long as the median redshift is sufficiently high ( ), Amara & Refregier (2007a) demonstrate that, it is more advantageous to perform the cosmic shear surveys in fields as wide as possible, rather than deep. This in order to minimize the error bars in cosmological parameters. To date, the largest data optimized for cosmic shear is the Wide field of the CFHTLS, which covers 50 deg^{2} (Fu et al. 2008) and will eventually reach 170 deg^{2}. Another analysis has also been published that combines 4 surveys that, together, cover an area of 100 deg^{2} (Benjamin et al. 2007). In a few years, the KIDS/VIKING survey will cover 1500 deg^{2}. Eventually, projects currently being planned, such as DUNE/Euclid, planned for 2017, will be able to perform cosmic shear measurements over the entire observable extragalactic sky ( deg^{2}).
To let cosmic shear surveys reaching their full potential, it is necessary to ensure that systematic errors are subdominant relative to statistical uncertainties. In particular, a tight control on all the effects associated with shape measurements is required. To illustrate the difficulty in compiling accurate shear measurements, we begin with an overview of the ``forward process'' that illustrates how the original image of a galaxy is distorted in forming the final image that we measure. In the forward process, a galaxy image is: (i) sheared by gravitational lensing; (ii) convolved with a point spread function (PSF) originating in a number of sources (e.g. instruments and atmosphere); (iii) pixelated at the detector; and finally (iv) affected by noise.
Cosmic shear analyses involve the reverse process: we begin with the final image and move backward from step (iv) to (i) in recovering the original lensing effect. A detailed and illustrated description of the forward and inverse processes is given in the GREAT08 Challenge Handbook (Bridle et al. 2008). The GREAT08 Challenge aims to provide a wide range of expertise into gravitational lensing by presenting the relevant issues in a clear way so as to make it accessible to a broad range of disciplines, including the machinelearning and signalprocessing communities. Other similar challenges have also been performed within the weak lensing community, as part of the STEP collaboration (Rhodes et al. 2008; Massey et al. 2007; Heymans et al. 2006), which focused mainly on understanding the systematic errors at play in current shear measurement methods. These challenges focus mainly on reducing the errors originating in the shape measurement method. However, even with a perfect method there are fundamental limits due the statistical potential of a data set.
In PaulinHenriksson et al. (2008) (P1 hereafter), we investigated the link between systematic errors in the power spectrum and uncertainties in the PSF correction phase. The framework is the following. Since the PSF of an instrument varies on all scales, the PSF needs to be measured using the stars that surround the lensed galaxy. Each star provides an image of the PSF that is pixelated and noisy, which means that to reach a given accuracy in the knowledge of the PSF, a number of stars is required. We estimate the number of stars N_{*} required to calibrate the PSF to a given accuracy, according to the stellar signaltonoise ratio (SNR), the minimum galaxy size, the complexity of the PSF and the tolerated variance in the systematics . On the other hand, Amara & Refregier (2007b) estimated the upper limit to when estimating cosmological parameters. By combining both papers together, we derive the minimum number of stars required to reach a given accuracy. For instance, analyses completed to date, that allow us to constrain with an accuracy of 0.05, require lower than a few 10^{6} and the PSF to be calibrated by using 5 stars; while for future ambitious surveys that will allow us to constrain w_{0} and w_{a} to an accuracy of 0.02 and 0.1, respectively, must be lower than 10^{7}, which requires at least 50 stars (for stars with signaltonoise ratio of 500 and a PSF described by a few degrees of freedom, as can be typically achieved in space).
In P1 we use the same functional form for both the underlying PSF as well as the model used to fit it. This means that the PSF model is able to describe perfectly the underlying PSF. The errors in the fit due to noise causes a scatter of the fitted parameters around the truth. For instance, if the model is an orthogonal basis set, then the fitted parameters follow a Gaussian distribution around the truth. In this paper, we extend this investigation by studying the impact of fitting a PSF with a model that has a different form. This addresses the case in which the underlying PSF (unknown in practice) is estimated by fitting the parameters of an arbitrary model. This can lead to both a scatter in the fitting parameters and an offset in the average value relative to the true value, i.e. a bias in the fitting parameters. We can therefore model a given PSF using either a complex model of small biases but large scatters, or a simpler model that would lead to smaller scatters but larger biases. To quantify these effects, we revisit the concept of complexity proposed in P1 and introduce the concept of sparsity.
This paper is organised as follow: first, in Sect. 2, we discuss the concepts of complexity and sparsity, which are the key concepts of this paper; Sect. 3 presents our notation; Sect. 4 presents the definition of optimal complexity, illustrates our formalism with a PSF example and uses the sparsity as a tool for optimizing the complexity; Sect. 5 derives the minimum number of stars required to calibrate the PSF, extending results of P1; and finally, Sect. 6 summarizes our conclusions.
2 Complexity and sparsity
In P1, we introduced the concept of complexity; we demonstrated that a few complexity factors characterize the amount of information that needs to be collected about the PSF. This is summarized and revisited in Sect. 2.1. In Sect. 2.2, we introduce the concept of sparsity, which measures the ability of a PSF model to represent the underlying PSF with a small number of free parameters. This allows us to explore how an optimal PSF model can be constructed to minimize .
2.1 Complexity
In P1, we define the complexity factors of the PSF, which represent the number of degrees of freedom (DoF) that are estimated from stars (in the limit of infinite resolution, i.e. infinitely small pixels): the higher the number of DoF, the larger the complexity factors.
Each PSF shape parameter is associated with a complexity factor that is related to the rms of its estimator.
In the simple formalism where we consider unweighted quadrupole moments, the PSF is characterized by only two complexity factors
and
associated with the 2 component PSF ellipticity
and the square PSF rms radius
respectively (as defined in P1).
For a given star, one has:
where is the photometric SNR of the star, is the rms error of the PSF size estimator and is the rms error of the PSF ellipticity estimator. As in P1, we assume that the small ellipticity regime holds (i.e. ), implying that the measure of is isotropic and the 2 components of the ellipticity have the same rms uncertainty, i.e. .
If the PSF can be considered constant over several stars, or for particular representations of the PSF (for example with shapelet basis sets in the small ellipticity regime, see P1),
and
are spatially constant and Eq. (1) can be extended to a set of several stars. For a combination of several stars,
becomes
:
where is the effective stellar SNR and n_{*} is the effective number of stars k used in the PSF calibration, as defined by (see P1):
We also show in P1 that the polar shapelet basis set, proposed by Massey & Refregier (2005), tested on simulated data in Massey et al. (2007) and used on real data by Bergé et al. (2008), is particularly convenient for modeling the PSF. For example, in the small ellipticity regime, and depend only on the polar shapelet basis set over which the PSF is decomposed, not on the PSF itself. For this reason we use shapelets in this paper when illustrating our discussions by an example. Note that our results and conclusions are not restricted to shapelets but remain valid whatever the PSF model. For convenience, we choose the shapelet ``diamond'' option (described in details in P1) that imposes a lower limit to the scales described, implying a link between and :
(4) 
with the highest even integer lower than or equal to the order of the basis set. We then consider the overall complexity defined in the following according to , and the variance in the galaxy ellipticity distribution.
Figure 1: PSF example (top panel) adopted in this paper and best fits of it (other panels) with 4 shapelet basis sets (corresponding to 2.6, 4.3, 7.8 and 16.4, i.e. 4, 6, 10 and 20 with the diamond option) and for equal to 100, 10^{3}, 10^{4} and (n_{*} is the numbers of stars used for the fit and is the effective SNR of stars, see Eq. (3); infinity corresponds to the ideal case of no background). For a given value of , all fits are performed with the same realization of the noise. Colors show the flux (darker colors indicate brighter regions) and show that this PSF is almost circular at the center. Contours show some isophotes not visible with the color scale and reveal the complex structure of the tails. The original (i.e. underlying) PSF in the top panel was built using a model with 24.8. The optimal value of the fitted complexity (in order to minimize ) is indicated under brackets for each value of . This figure illustrates that, for a given value of , the simpler the model (i.e. the lower ), the poorer the description of the tails and the larger the bias. On the other hand, for a given , the lower the amount of information available in stars (i.e. the smaller ), the noisier the description of the tails. 

Open with DEXTER 
2.2 Sparsity
In this paper, we introduce the concept of sparsity of the PSF model, which describes how efficiently a model can represent the underlying PSF with a limited number of DoF (i.e. with a limited complexity). Specifically, the sparsity quantifies how the residuals between the estimated and the underlying PSF decrease as the complexity of the PSF model increases. With a high number of DoF, i.e. a high complexity, one might expect small residuals but large scatters ointhe fitted parameters. On the other hand, with a small number of DoF, i.e. a low complexity, one might expect large residuals but small scatters in the fitted parameters. The sparsity characterizes the slope in this relation and thus is an estimate of the amount of information that can be contained in a given number of DoF. We show how to use sparsity in optimizing the complexity and minimizing .
Consider the shape parameters
and
of the underlying PSF, as defined previously.
The differences
and
between the underlying PSF (``true'' index) and its estimation (``est'' index) can be written:
(5) 
These differences are of two types: the statistical scatter relative to the average and the biasoffset b of the average relative to the true value. The mean square errors (MSE) of and are:
In P1, we address the zero bias case , that is equivalent to consider the PSF model is able to describe the underlying PSF perfectly. However, this nulling of the biases is not necessarily the optimal PSF modeling. It can be advantageous to work with a simplistic PSF model that is unable to describe all the PSF features and has some biases but low statistical scatter (see our PSF example in Sect. 4.1 and Fig. 1). This paper proposes another approach that consists in optimizing the PSF model in order to minimize . We do this by searching for the optimal tradeoff between the systematic errors ( , , and b[R^{2}]) and the statistical errors ( , and ), which is equivalent to searching the optimal complexity of the PSF model. The sparsity allows us to perfom this optimization because it characterizes the decrease in the biases b as increases.
In the following, we define a ``sparsity parameter'' in the particular case where the biases are modeled as a powerlaw function of the PSF model complexity ( ), and we study the impact of on the number of stars required to calibrate the PSF. We thus revisit the main result of P1 by deriving N_{*} (the number of stars required to calibrate the PSF) according to instead of (the complexity of the underlying PSF). Moreover, this new relation is optimized to minimize .
We emphasize that, in this paper, we propose to optimize the complexity of the PSF model within a given the basis set. We do not address the issue of choosing the basis set itself. There is no doubt that, to optimize the PSF modeling, it is necessary to select carefully this basis set. For instance, generic basis sets such as shapelets, wavelets, or Fourier modes, although they have enormous advantages, are not optimal. This issue will be addressed in forthcoming works.
3 PSF calibration for shear measurement
When deconvolving the observed galaxy with the estimated PSF,
and
propagate into an
error
in the estimation of the galaxy ellipticity. We denote
and
,
the rms radius and the twocomponent ellipticity of the galaxy. When
,
,
,
and
are defined using the unweighted moments of the flux, this propagates to (see P1):
The spatial average of is related to the variance in the systematic errors in the shear measurements (Amara & Refregier 2007b), defined by the integral of the systematics in the power spectrum:
with the calibration factor between the gravitational shear and the ellipticity. Its value depends on the distribution of galaxy ellipticities and is typically about 1.84 (Rhodes et al. 2000). The brackets < > denote a spatial average over the entire field. As in P1, we substitute Eq. (8) into Eq. (10) with the following simplifying assumptions:
 1.
 The galaxy is not correlated with the PSF.
 2.
 The error on the PSF ellipticity ( ) and the PSF ellipticity itself ( ) are not correlated. This is warranted by the fact that, in the assumed small ellipticity regime, does not have any preferred direction, implying that 0.
 3.
 We neglect correlations between the ellipticity and the inverse squared radius of the galaxy. This is reasonable for the PSF calibration in the small ellipticity regime.
We develop a more compact expression by adopting the following notation:
which leads to:
In P1, we considered only the scatters and (i.e. in the zero bias case: ) and we approximate the statistical averages with spatial averages: and . In this paper, with the introduction of biases, the scatter becomes the MSE:
This leads to:
We can see that is proportional to the quadratic sum of 6 terms: three bias terms and three statistical ones. Collecting terms of similar type using the following notation:
gives:
Although only depends on the SNR of the stars, B and both depend on the complexity of the modeling. They can be denoted and . In P1, we show that it is given by:
where the overall complexity of the model is given by the complexities and associated with the ellipticity and the squared radius of the model, respectively (see Eq. (2)):
Equations (19) and (20) then infer that:
From this equation, we can see that increasing the complexity by adding degrees of freedom in the PSF model can reduce B but also increases the statistical errors. Minimizing thus implies the search for the optimal tradeoff in the value of .
4 Optimal PSF model
In Sect. 4.1, we present a PSF example that we use in the remainder of this paper to illustrate our discussion. In Sect. 4.2, we show the optimal complexity of the PSF model (that minimizes ) and apply this to the PSF example. We then explore this optimization in more details in Sect. 4.3 by examining a particular case in which the bias can be described by a powerlaw function of the complexity.
4.1 PSF example
To illustrate our discussion, we study a realistic example of a PSF with complex features in the tails, and investigate what happens when fitting it with various shapelet basis sets as function of the SNR of the available stars. We also use a shapelet basis set (which differs from that used in the fits) for describing the underlying PSF. This use of shapelets for both the PSF model and the underlying PSF was chosen for three reasons:
 
 first, it allows pixelation issues to be ignored, which are beyond the scope of this paper. Indeed, the description of the underlying PSF is performed by using the continuous shapelet functions and the fits are performed at high resolution;
 
 second, it considerably simplifies both the calculations and the fitting process, due to the orthogonality of shapelet functions (the average estimation of a fitted coefficient is the true value, independently of the other coefficients);
 
 third, it is a simple and convenient framework for illustrating the use of sparsity as a tool in optimizing the complexity of the PSF modeling.
In Fig. 1, we also show the 16 fits performed with the 4 shapelet basis sets (corresponding to , 6, 10, and 20) and (the stellar signaltonoise ratio, see Eq. (3)) equal to 100, 10^{3}, 10^{4}, or (the latter is the ideal case of no background). To determine the overall complexity , which depends on the rms of galaxy ellipticities in terms of the parameter (see Eqs. (13) and (21)), we adopt the typical value , for which , 6, 10, 20, 34 correspond to , 4.3, 7.8, 16.4, 28.4 respectively. In the following, we also adopt the value , that corresponds to the typical values and . Figure 1 illustrates that:
 
 when is sufficiently high, complex basis sets are required to model the complex tails, i.e. the amount of bias B decreases as the complexity of the model increases. For instance, a fit with and would allow one to recover our PSF example exactly, with B=0;
 
 a higher complexity requires a higher number of DoF to be fitted. Consequently, for a given value of , increasing the complexity of the model also increases the scatter in the estimated shape. Therefore, it is not always appropriate to use a complex fit model; it may be more robust to use a simplified (but more biased) fit model.
4.2 Optimizing the complexity of the PSF model
The optimal PSF model is that for which
is minimized, varying .
We define the optimal value
to be the minimal value of
and in the same spirit, we note that
is the corresponding value of :
For instance, Fig. 2 illustrates the search for the optimal shapelet basis when our PSF example (see the previous section) is estimated with 50 stars (and with ). The optimal model is that corresponding to (i.e. with the diamond option) and , shown by the red diamond.
Figure 2: Total variance and its contributions (see Eq. (16)) with respect to the fit model complexity , for our PSF example (see Sect. 4.1 and Fig. 1) fitted with 50 stars (i.e. n_{*}=50), and . We note and . The optimal basis set is that for which is minimum. At this point, shown by the red diamond, and . 

Open with DEXTER 
For a given fit model, increasing n_{*} reduces the scatters but not the biases in the model fitting (see Eq. (22)). Therefore, as n_{*} increases, increases and decreases. This is illustrated in Fig. 3, which shows B, , and (see Eqs. (17) to (19)) when our PSF example is estimated with n_{*}=10, 50, or 200 (still ).
Figure 3: Total variance (3 thick and solid colored curves) and its contributions to B (one thin and solid black curve) and (3 dashed colored curves), as defined by Eqs. (17) to (19), with respect to the PSF model complexity , for our PSF example (presented in Sect. 4.1 and Fig. 1) fitted with 10 (red curves), 50 (green curves) or 200 (blue curves) stars (i.e. n_{*}=10, 50, 200) with an effective SNR of . depends on n_{*} (see Eq. (20)), implying a different curve for each value of n_{*}, while B does not. This illustrates that (i.e. the minimum value of shown by the diamonds) increases with n_{*}. 

Open with DEXTER 
Figure 4 shows as a function of n_{*}. The diamonds represent the curve for our PSF example illustrated in all previous plots, while the boldstraight line without any diamond shows the ideal case (addressed in P1) of a PSF described perfectly by the model (i.e. B=0). Thus, varies with 1/n_{*} as predicted by our scaling relation presented in P1. The dotted and dashed lines are discussed in Sect. 4.3.
Figure 4: Optimal variance of the shape measurement systematics as a function of the number of stars n_{*} used to calibrate the PSF. The diamonds show the curve for our PSF example illustrated in all previous plots (presented in Fig. 1) and fitted with shapelets. The straight bold line shows the ideal case (addressed in P1) of a PSF model that exactly describes the PSF (i.e. with no residual). The horizontal line shows the values which is the requirement to be able to constraint w_{0} and w_{a} at 0.02 and 0.1 respectively (Amara & Refregier 2007b). The blue lines (dashed, dotted and dotteddashed) are discussed in Sect. 4.3. They are the curves expected when modeling the bias B with a powerlaw function of the complexity as stated by Eq. (24) and illustrated in Fig. 5. 

Open with DEXTER 
4.3 Example of optimal complexity in the case of a powerlaw function
In this section, we derive the optimal complexity when the bias B is a powerlaw function of the complexity written as
.
We investigate
as a function of n_{*} and .
We normalize the powerlaw function, such as:
In our example of a PSF fitted with a shapelet basis set (see Sect. 4.1 and Fig 1), the smallest value of that we consider is 2.6 (which corresponds to with the ``diamond'' configuration, see P1). This explains why, for this example, we choose to normalize the powerlaw function to this value , implying . This model is illustrated in Fig. 5 for equal to 2, 4, 6, and 8. This is superimposed on the B versus relation that we obtain when fitting our PSF example with different shapelet basissets, as decribed in Sect. 4.1. We see that in this case B is reasonably described by .
This representation by a powerlaw function is particularly convenient because can be identified with the sparsity: a high value of means that the PSF model is efficient in representing the underlying PSF with a small number of free parameters. Conversely, a low value of means the PSF model requires a large number of parameters to describe the underlying PSF without large residuals. In the following, is called the ``sparsity parameter''. Together with this powerlaw representation (Eq. (24)), Eqs. (22) and (23) imply that:
Note that Eq. (27) expresses (the minimum variance in the systematic errors in shear measurements that can be achieved, see Eqs. (9), (10), and (23)) in terms of a set of parameters that can be divided into 2 families:
 1.
 parameters that are properties of the data set, such as , B_{0}, , and ;
 2.
 parameters that are properties of the analysis method, such as n_{*} (i.e. the number of stars used to calibrate the PSF) and (i.e. the sparsity parameter of the PSF model).
Figure 5: Overall bias B versus for our PSF example (in black) and some powerlaw functions , , 4, 6, 8 (in blue) normalized to intercept at . We see that the case fits well with the example. 

Open with DEXTER 
5 Required number of stars
As discussed in the introduction, an important issue for cosmic shear surveys is to ensure that systematics are kept smaller than the statistical errors, by demanding an upper limit to . Part of the systematics have their origin in the PSF calibration, which is imperfect due to the limited number of stars available. In this section, we express N_{*}, the number of stars required to calibrate the PSF, in terms of the level of systematic errors (note the capital ``N'', as opposed to ``n_{*}'' which is the number of stars involved in the PSF calibration process: we need to ensure that systematic effects are below . N_{*} is the lower limit of n_{*}).
In Sect. 5.1, we summarize the conclusions of P1 which apply when the underlying PSF and the PSF model have the same functional form (i.e. B=0) and we extend these conclusions to the general case of PSF modeling performed with any model (i.e. B not necessarily equal to 0). In Sect. 5.2, we invert Eq. (27) (that holds when B is described by a powerlaw function of the complexity: ) and express N_{*} as a function of and of the minimum systematic level achievable when the complexity of the PSF modeling is optimal.
5.1 Generalised scaling relation
In the optimistic case where the PSF calibration is the only significant source of systematic errors, a given value of N_{*} (i.e. a given number of stars involved in the PSF calibration) implies a value of
.
This is presented in P1 in the form of a scaling relation that links N_{*},
,
(the effective signaltonoise ratio of stars),
(the ratio between the smallest galaxy size and the PSF size), and
(the complexity of the PSF):
The factor 2 at the end comes from the fact that (in P1, this scaling relation is written in terms of ). This holds for the assumption that the PSF model is able to describe the PSF without any bias (i.e. B=0). With nonzero B and adopting the same simplifications and the same typical values as in P1, Eq. (22) leads to the more general relation:
where
Thus, taking B into account in the scaling relation translates into the new factor in Eq. (30), which equals 1 when B is zero (then the relation (29) is equivalent to the scaling relation given in P1) and is related to the ratio , which is the relative weight of biases in the error budget.
Figure 6: as defined in Eq. (32). 

Open with DEXTER 
5.2 Application to the powerlaw model
Equation (27) can be inverted to provide N_{*} (the number of stars required to calibrate the PSF) as a function of
(the minimum level of systematics achievable when optimizing the complexity of the PSF modeling),
(the sparsity parameter),
(the effective SNR of stars defined in Eq. (3)), and
(a dimensionless factor defined in Eq. (12)):
with the dimensionless function:
shown in Fig. 6. With the notation and scaling of Eq. (29), Eq. (31) is equivalent to:
This equation allows one to estimate the number of stars required to calibrate the PSF and thus, with respect to the stellar density, the minimum scale on which the PSF calibration is possible. On smaller scales, stars provide insufficient information for calibrating PSF variations. This implies that these smaller scales may be contaminated by systematics due to a poor correction of the PSF and should not be used to estimate cosmological parameters, unless the variabilities on small scales are known to be extremely small. As shown by Amara & Refregier (2007b) and discussed in P1, future allsky cosmic shear surveys will need to achieve to be able to estimate w_{0} and w_{a} with uncertainties of about 0.02 and 0.1, respectively. Fig. 4 shows that, for our PSF example, it is possible to achieve this accuracy when calibrating the PSF with 50 stars, if . Although this is not a general statement (this assumes that B can be described by a powerlaw function, see Eq. (24), and depends on the normalization parameter B_{0}), this is a representative example of the sparsity requirement for future cosmic shear surveys. On the other hand, for current cosmic shear surveys of areas deg^{2}, we have the requirement (see Amara & Refregier (2007b) and P1). In this case, it is possible to calibrate the PSF with a few stars when . This sparsity requirement is reached with the current PSF correction methods (for instance that based on shapelets as in Bergé et al. 2008), and, assuming a star density of about 1 per arcmin^{2}, this is consistent with the presence of significant B modes usually found on scales smaller than a few arcmins.
6 Conclusions
We explore the systematics induced in cosmic shear by the PSF calibration/correction process and study how to optimize the PSF model to minimize the systematic errors in cosmological parameter estimations. In this framework, we revisit the concept of the complexity of the PSF, defined in our previous paper (P1), and introduce the concept of the ``sparsity'' of the PSF model. The complexity characterizes the number of degrees of freedom in the model. A small number of degrees of freedom corresponds to a low and relates to a simple PSF model, which can be fitted to the stellar observations of low signaltonoise ratio, but is likely to be highly biased. On the other hand, a large number of degrees of freedom corresponds to a high and relates to a complex PSF model, which is expected to have a low bias but requires stellar observations of high signaltonoise ratio to avoid large statistical scatters in the fitted parameters. In P1, we related the complexity of the PSF model to the systematic errors in cosmological parameter estimations. In this paper, we show how the complexity can be optimized depending on the stars available, using the concept of sparsity. The sparsity characterizes the decrease of residuals between the best fit of the PSF model on the underlying PSF, when adding degrees of freedom to the model.
In the general case, we also extend the scaling relation, proposed in P1, between the number of stars used to calibrate the PSF and the systematic errors in the cosmological parameter estimations.
As discussed in P1, this relation, with the constraint of maintaining the systematics below the statistical uncertainties when estimating cosmological parameters, infers the number of stars N_{*} required for the PSF calibration. N_{*} corresponds to the minimum scale on which the PSF modeling is accurate: on scales smaller than this minimum, there is insufficent information in the data to calibrate the PSF variations. This implies that these smaller scales may be contaminated by systematics related to a poor PSF correction and should not be used when estimating cosmological parameters (unless the variabilities are known to be small, due, for instance, to the quality of the hardware).
We consider a realistic PSF example and model the amount of bias B between the PSF fit and the underlying PSF by a powerlaw function of the fitted complexity: where is the sparsity parameter. We find that, for this PSF, current cosmic shear analyses that cover or less, need to be higher than 2, which is achievable by current analysis methods. Thus, current cosmic shear analyses do not require a rigorous optimization of the PSF model. On the other hand, future cosmic shear surveys that aim to measure w_{0} and w_{a} to an accuracy of 0.02 and 0.1, respectively, will require to calibrate the PSF with 50 stars. This relation between the required number of stars N_{*} and the accuracy of the calibration depends on the underlying PSF. This explains why these values, although corresponding to realistic orders of magnitude, cannot be assumed to represent a general result. Two parameters drive this relation: the amount of biases B_{0} when fitting the underlying PSF with a PSF model of low complexity (N_{*} being proportional to ; in our example, ), and the sparsity parameter of the PSF modeling during the analysis.
It is thus possible to optimize cosmic shear surveys at two levels: when optimizing the observational conditions, the PSF must be as simple and stable as possible in order to make possible its description by a low complexity model (this minimizes B_{0}); when analysing the data, the PSF modeling must be optimized to have as high a value of the sparsity as possible.
The approach suggested in this paper is a first step toward introducing the concept of sparsity to weak lensing shape measurements. We do not address issues related to the pixelation. Moreover, although we only address the PSF calibration, this approach is also applicable to other topics such as description of galaxy shapes.
Acknowledgements
We thank JeanLuc Starck and Jérome Bobin for useful discussions and insight on sparsity. We also thank Sarah Bridle and Lisa Voigt for an ongoing collaboration on weak lensing shape measurements. SPH is supported by the P2I program, contract number 102759. AA is supported by the Swiss Institute of Technology through a Zwicky Prize.
References
 Amara, A., & Refregier, A. 2007a, MNRAS, 381, 1018 [NASA ADS] [CrossRef] (In the text)
 Amara, A., & Refregier, A. 2007b, [arXiv:0710.5171] (In the text)
 Benjamin, J., Heymans, C., Semboloni, E., et al. 2007, MNRAS, 381, 702 [NASA ADS] [CrossRef] (In the text)
 Bergé, J., Pacaud, F., Réfrégier, A., et al. 2008, MNRAS, 385, 695 [NASA ADS] [CrossRef] (In the text)
 Bridle, S., ShaweTaylor, J., Amara, A., et al. 2008, [arXiv:0802.1214] (In the text)
 Fu, L., Semboloni, E., Hoekstra, H., et al. 2008, A&A, 479, 9 [NASA ADS] [CrossRef] [EDP Sciences]
 Heymans, C., VanWaerbeke, L., Bacon, D., et al. 2006, MNRAS, 139, 313
 Massey, R., & Refregier, A. 2005, MNRAS, 363, 197 [NASA ADS] [CrossRef] (In the text)
 Massey, R., Heymans, C., Bergé, J., et al. 2007, MNRAS, 376, 13 [NASA ADS] [CrossRef]
 PaulinHenriksson, S., Amara, A., Voigt, L., Refregier, A., & Bridle, S. L. 2008, A&A, 484, 67 [NASA ADS] [CrossRef] [EDP Sciences]
 Rhodes, J., Refregier, A., & Groth, E. J. 2000, ApJ, 536, 79 [NASA ADS] [CrossRef] (In the text)
 Rhodes, J., Refregier, A., & Groth, E. J. 2008, in preparation
Footnotes
 ... Survey^{}
 http://www.cfht.hawaii.edu/Science/CFHLS/
 ... Survey^{}
 http://www.eso.org/sci/observing/policies/PublicSurveys/sciencePublicSurveys.html
 ... Survey^{}
 http://www.darkenergysurvey.org
 ... System^{}
 http://panstarrs.ifa.hawaii.edu
 ... Probe^{}
 http://snap.lbl.gov
 ... Telescope^{}
 http://www.lsst.org
 ... Explorer^{}
 http://www.dunemission.net and http://www.esa.int/esaCP/index.html
All Figures
Figure 1: PSF example (top panel) adopted in this paper and best fits of it (other panels) with 4 shapelet basis sets (corresponding to 2.6, 4.3, 7.8 and 16.4, i.e. 4, 6, 10 and 20 with the diamond option) and for equal to 100, 10^{3}, 10^{4} and (n_{*} is the numbers of stars used for the fit and is the effective SNR of stars, see Eq. (3); infinity corresponds to the ideal case of no background). For a given value of , all fits are performed with the same realization of the noise. Colors show the flux (darker colors indicate brighter regions) and show that this PSF is almost circular at the center. Contours show some isophotes not visible with the color scale and reveal the complex structure of the tails. The original (i.e. underlying) PSF in the top panel was built using a model with 24.8. The optimal value of the fitted complexity (in order to minimize ) is indicated under brackets for each value of . This figure illustrates that, for a given value of , the simpler the model (i.e. the lower ), the poorer the description of the tails and the larger the bias. On the other hand, for a given , the lower the amount of information available in stars (i.e. the smaller ), the noisier the description of the tails. 

Open with DEXTER  
In the text 
Figure 2: Total variance and its contributions (see Eq. (16)) with respect to the fit model complexity , for our PSF example (see Sect. 4.1 and Fig. 1) fitted with 50 stars (i.e. n_{*}=50), and . We note and . The optimal basis set is that for which is minimum. At this point, shown by the red diamond, and . 

Open with DEXTER  
In the text 
Figure 3: Total variance (3 thick and solid colored curves) and its contributions to B (one thin and solid black curve) and (3 dashed colored curves), as defined by Eqs. (17) to (19), with respect to the PSF model complexity , for our PSF example (presented in Sect. 4.1 and Fig. 1) fitted with 10 (red curves), 50 (green curves) or 200 (blue curves) stars (i.e. n_{*}=10, 50, 200) with an effective SNR of . depends on n_{*} (see Eq. (20)), implying a different curve for each value of n_{*}, while B does not. This illustrates that (i.e. the minimum value of shown by the diamonds) increases with n_{*}. 

Open with DEXTER  
In the text 
Figure 4: Optimal variance of the shape measurement systematics as a function of the number of stars n_{*} used to calibrate the PSF. The diamonds show the curve for our PSF example illustrated in all previous plots (presented in Fig. 1) and fitted with shapelets. The straight bold line shows the ideal case (addressed in P1) of a PSF model that exactly describes the PSF (i.e. with no residual). The horizontal line shows the values which is the requirement to be able to constraint w_{0} and w_{a} at 0.02 and 0.1 respectively (Amara & Refregier 2007b). The blue lines (dashed, dotted and dotteddashed) are discussed in Sect. 4.3. They are the curves expected when modeling the bias B with a powerlaw function of the complexity as stated by Eq. (24) and illustrated in Fig. 5. 

Open with DEXTER  
In the text 
Figure 5: Overall bias B versus for our PSF example (in black) and some powerlaw functions , , 4, 6, 8 (in blue) normalized to intercept at . We see that the case fits well with the example. 

Open with DEXTER  
In the text 
Figure 6: as defined in Eq. (32). 

Open with DEXTER  
In the text 
Copyright ESO 2009