A&A 414, 17-21 (2004)
DOI: 10.1051/0004-6361:20031632
R. Vio1 - P. Andreani2 - W. Wamsteker3
1 - Chip Computers Consulting s.r.l., Viale Don L. Sturzo 82,
S. Liberale di Marcon, 30020 Venice, Italy
ESA-VILSPA, Apartado 50727, 28080 Madrid, Spain
2 -
Osservatorio Astronomico di Padova, vicolo dell' Osservatorio 5,
35122 Padua, Italy
3 -
ESA-VILSPA, Apartado 50727, 28080 Madrid, Spain
Received 15 August 2003 / Accepted 16 October 2003
Abstract
In this paper we comment on the results concerning the performances of matched filters, scale
adaptive filters and Mexican hat wavelet that recently appeared in literature in the context of
point source detection in Cosmic Microwave Background maps. In particular, we show that,
contrary to what has been claimed, the use of the matched filters still appear to be the most reliable
and efficient method to disantangle point sources from the backgrounds, even when using
detection criterion that, differently from the classic
thresholding rule, takes into account
not only the height of the peaks in the signal corresponding to the candidate sources but also their curvature.
Key words: methods: data analysis - methods: statistical - cosmology: cosmic microwave background
Studying diffuse backgrounds in all-sky maps implies the possibility of disentangling background signals from those originated from point sources. This task is of fundamental importance in dealing with Cosmic Microwave Background (CMB) data. In this context various papers studied the "optimal'' method for such a task. Three main methods have been considered so far: the Mexican hat wavelet (Cayon et al. 2000), the scale-adaptive filters (or optimal pseudo-filters) and the matched filters (Sanz et al. 2001; Vio et al. 2002 and reference therein). Matched filter (MF) is constructed taking into account the source profile and the background to get the maximum signal-to-noise ratio (SNR) at the source position. Scale-adaptive filter (SAF) is built similarly to MF with the additional constraint to have a maximum in filtered space at the scale and source position. The Mexican hat wavelet (MHW) represents a separate case since it is "a priori'' filter, adapted to the detection of point sources. Its main limitation is that it is founded on semi-empirical arguments and therefore lacks a rigorous theoretical justification. For this reason, in the following we will be especially concerned with MF and SAF.
Vio et al. (2002) (henceforth VTW) have shown that, in spite the claims of "optimality'' for SAF and MHW (Sanz et al. 2001, henceforth SHM), in reality these filters do not behave as good as the MF. In a recent work in the context of one-dimensional signals, Barreiro et al. (2003, henceforth BSHM) compare SAF, MHW, and MF on the basis of a detection criterion based on the Neyman-Pearson decision rule, that takes into account not only the height of signal peaks but also their curvature. These authors find that, although MF is effectively optimal in most of the cases, there are situations where SAF and MHW can overperform it. Here we show that such a result is not correct since it is linked to the measure of performance adopted by authors, that tends to favour the filters characterized by a low detection capability. MF is in general superior to these other two filters.
For sake of generality, we firstly present our arguments in Rn and then we specialize the results to the one-dimensional case.
The sources are assumed to be point-like signals convolved with the beam of the measuring instrument and are
thus assumed to have a profile
.
The signal
,
,
is modeled as
![]() |
(2) |
![]() |
(4) |
The classic procedure for the detection of the sources consists in filtering signal to enhance
the sources with respect to the background. This is done by cross-correlating the signal
with a filter
.
The source locations are then determined by selecting the peaks in the filtered signal
that are above a chosen threshold. Finally, the source amplitudes are estimated as the
values of the filtered signal at the estimated locations. The question is the selection of an
optimal filter
for such procedure. In order to define it, some assumptions are necessary.
In particular it is assumed that the source profile and background spectrum are known,
the profile is spherically symmetric, characterized by a scale
,
and the background is
isotropic. These assumptions allow to write
,
where
,
and
for
.
In addition,
source overlap is assumed negligible. In the present context, we are interested in
the general family of spherically symmetric filters
of the form
.
The cross-correlation between
and
provides a filtered
field
with mean
and variance
.
Source locations are assumed to be known
and the aim is to estimate the amplitudes. Given the assumed distance between the sources, it is enough to
consider
a field
as in Eq. (1) with a single source at the origin,
.
Its amplitude
is estimated by requiring it to be an unbiased estimator of A, i.e.,
.
On the other hand, to enhance the magnitude of the source relative to the background
the filter
is required to minimize
the variance
.
This has the effect of maximizing, among unbiased estimators,
the detection level
Since
is chosen in a way that
is a minimum variance linear (in
)
unbiased
estimator of A, it follows that (Gauss-Markov theorem)
is the (generalized) least squares estimate
of A achieved by the filter
In the pseudo-filter approach of SHM the filters have the same form
of
with an additional scale
dependence
To determine an optimal filter ,
SHM minimize the variance of the filtered field
with the two constraints:
is required to be, as in the previous section,
an unbiased estimator of A for some known
,
and
is selected so
that
has a local maximum at scale R0. This latter translates into
b | ![]() |
![]() |
(11) |
c | ![]() |
![]() |
(12) |
In their work VTW stress the fact that, since both
and
provide a
linear and unbiased estimate of
the amplitude A then, regardless the source profile and background spectrum
and because of the optimality of the least squares,
.
As a consequence the value of the detection level,
,
corresponding
to
is at least as high, or higher, than that achieved with
.
Furthermore, via an extensive
set of numerical simulation VTW have shown that this conclusion holds even when the source
location uncertainty is taken into account. In other words, enough information
about the scale of the source is already included in the derivation of the matched filter.
Via numerical simulations VTW have also shown that MF overperforms SAF when comparing the resulting
numbers of incorrectly detected sources. VTW's conclusion is then nothing is gained
by using SAF.
Recently, in the context of one-dimensional signals, zero-mean Gaussian background with scale-free power spectrum
,
and Gaussian profile
for the source, BSHM
criticized
these conclusions through the argument that the detection level
and the
thresholding method used by VTW as detection rule are not sufficient to support
their results
. For this reason, they introduce a new detection criterion based on a Neyman-Pearson
decision rule which uses not only the heigth of the maxima in the signal but also their curvature. This
method can be summarized as follows (for more details, see BSHM)
If the 1D background z(x) is Gaussian, then it is possible to estimate the expected total number density nb
of maxima (i.e., number of maxima per unit interval in x) as well their expected number density
per intervals
and
,
where
and
are the normalized field and
curvature, respectively. Here,
is the moment of order 2n associated with the field.
If all the sources are assumed to have the same amplitude A,
it is possible to estimate the corresponding quantities n and
,
,
when the sources are embedded in the background. These quantities
allow to calculate, for any region
,
the probability density functions
![]() |
(14) |
In order to obtain a detection criterion, BSHM introduce the significance s2
![]() |
(17) |
![]() |
(18) |
It happens that for SAF, MF, and MHW, and independently from the index ,
s2 is maximized for
.
Figure 1 shows the corresponding R* for sources with an amplitude A such as
after filtering
with SAF. This figure shows that, at variance with SAF and MHW, the acceptance region of MF
does not depend on the curvature
but only on the height of the maxima. Therefore, for MF the detection rule
proposed by BSHM provides a criterion similar to the classic
thresholding rule.
Once fixed R* it is possible to calculate the expected number density nb* of incorrect and the expected
number density n* of correct detections
by integrating
and
over R*. These quantities are used by BSHM
to calculate the ratio r=n*/nb, called reliability, and the quantity
![]() |
(20) |
![]() |
Figure 1:
Acceptance region R*, when
![]() ![]() ![]() ![]() |
Open with DEXTER |
First, similarly to VTW and in spite of the introduction of the
new detection criterion, BSHM find that in most situations
the use of the second constraint (9) in SAF is not only useless but even harmful.
Second, if, on the one hand, the superiority of MF for
and
is out of discussion (this filter provides the largest number of correct detections and the smallest number
of incorrect ones),
the same conclusion for SAF when
is questionable. In this range of
SAF provides a smaller number of incorrect detections, but at the same time also a
smaller number of correct ones.
In this respect, at least in principle, the reliability parameter r should be
used as a measure of the filter performances only when an incorrect detection has a larger "cost'' than
missing a source, a fact that has to be proved in the context of CMB. Furthermore, even in the case of an
high "cost'' for the incorrect detections, r has to be used with great care.
The reason is that MF is constructed in such a way to maximize source detections. Therefore, the maximization
of s2 with respect to L* provides a criterion favouring the detection of a true source rather than the
rejection of a false one.
If one is worried of incorrect detections, there is a simple cure: the choice of
a L* making the detection of the sources less efficient.
In this way, part of the correct detections
will be lost but also the number of incorrect detections will decrease. Furthermore, in case of sources embedded
in the background the signal peaks are expected to have a mean height larger than that expected in case of only
background signal. Therefore, the smaller detection efficiency will affect more the number of incorrect detections
than that
of the correct ones. This fact is shown in Fig. 3 where it is evident that, when
and
,
MF has the same number density of incorrect detections as SAF with L*=1 but still a larger
number density of correct detections and consequently a larger
reliability r. The conclusion is that, as done in VTW, a meaningful evaluation of the performances of the two
filters requires that the comparison is made by fixing the number density of incorrect (or alternatively, correct)
detections.
If n*b is set at the value of SAF for L* = 1, the quantity r shown in Fig. 4 indicates that
MF is better than SAF also for
.
Similar arguments hold also for MHW that BSHM claim to provide a slightly better performance than MF
when
.
Figure 5 shows again that this conclusion is not correct.
The arguments presented in the previous section have been developed under the hypothesis that all the sources
are characterized by the same amplitude A. Of course, this condition is not satisfied in real situations.
In order to solve this problem, BSHM suggest to substitute the likelihood ratio (19)
with:
![]() |
(21) |
![]() |
(22) |
The first, and most obvious, is that such a conclusion suffers the same limitation
found in the previous section. Consequently the claim of superiority of SAF and MHW over MF is again not founded.
The second comment is that, in order to obtain reliable results,
is needed to
be known with good accuracy: the use of a wrong
will end in a false rule according to which
overweighs the smallest amplitudes or the largest ones, favouring
the (correct and incorrect) detections or the (correct and incorrect) rejections with obvious consequences
on the "optimality'' of the method.
In the framework of CMB studies the a priori information on
is not
available or is very inaccurate. The consequence is that a simple detection rule as, for example, the
thresholding criterion
could still represent the best choice since it requires the only a priori knowledge of the noise level. This
approach is
much simpler and safer than estimating the distribution of the source amplitudes.
![]() |
Figure 2:
Relationship ![]() |
Open with DEXTER |
![]() |
Figure 3:
Relationship L* vs. the number density nb* of incorrect detections, the
number density n* of correct detections, the
reliability r, and the relative detection ratio D corresponding to MF for
![]() |
Open with DEXTER |
![]() |
Figure 4:
Relationship between ![]() |
Open with DEXTER |
![]() |
Figure 5:
Relationship ![]() |
Open with DEXTER |
This paper deals with the detection techniques to extract point-sources from Cosmic Microwave Background maps. Various recent works appeared in the literature, presenting new techniques with the aim to improve the performances of the classical matched filters (MF). In particular the scale adaptive filters (SAF) and the Mexican hat wavelet (MHW) have been proposed as the most efficient and reliable methods (see Sanz et al. 2001, and references therein). This claim was subject to criticism by Vio et al. (2002) since they showed that in reality SAF and MHW have performances that in general are inferior to those provided by MF.
Recently Barreiro (2003) used the argument that
a criterion making use of a simple
thresholding rule is not fully sufficient to claim
detection. To support this assertion Barreiro (2003), in the context of one-dimensional signals and sources
with Gaussian profiles, adopt a
detection criterion based on a Neyman-Pearson decision rule that makes use of both the height and the curvature
of the maxima in the signal. Their theoretical arguments and numerical simulations indicate that, although
in general MF still remains the filter with the best performances, there are situations where SAF and
MHW overperform it.
In this paper we show that this conclusion is again not correct since
it is basically founded on a performance test favouring the filters characterized by a low
detection capability. This means that there is no reason to prefer SAF or MHW to MF. Furthermore, the claimed superiority
of SAF and MHW, when the source scale has to be estimated from the data, has still to be
proved, and in principle also MF could be modified in such a way to efficiently deal with this situation.
These conclusions are not academic: the use of non-standard statistical tools
is indicated only in situations
of real and sensible improvements of the results. New techniques that do not fulfill this requirement should
be introduced with care: they prevent the comparison with the results obtained in other works and may lead
people to use not well tested methodologies (MF has been successfully used for many years in very
different scientific contextes) ending up in not reliable results. Moreover, in the
present context, the use of SAF introduces further complications in the
analytical form of the filters (e.g., compare Eq. (6) with Eq. (10)) and in the
definition of the detection rule (for MF the calculation of the curvature
of the peaks in the signal
is not required).