Issue |
A&A
Volume 494, Number 2, February I 2009
|
|
---|---|---|
Page(s) | 707 - 717 | |
Section | Planets and planetary systems | |
DOI | https://doi.org/10.1051/0004-6361:200810754 | |
Published online | 11 December 2008 |
Investigating the potential of the Pan-Planets project using Monte Carlo simulations
J. Koppenhoefer1,2 - C. Afonso3 - R. P. Saglia1,2 - Th. Henning3
1 - University Observatory Munich, Scheinerstrasse 1, 81679
München, Germany
2 - Max
Planck Institute for Extraterrestrial Physics, Giessenbachstrasse,
85748 Garching, Germany
3 - Max Planck Institute for Astronomy,
Königstuhl 17, 69117 Heidelberg, Germany
Received 6 August 2008 / Accepted 2 December 2008
Abstract
Using Monte Carlo simulations we analyze the potential of
the upcoming transit survey Pan-Planets. The analysis covers the
simulation of realistic light curves (including the effects of
ingress/egress and limb-darkening) with both correlated and
uncorrelated noise as well as the application of a
box-fitting-least-squares detection algorithm. In this work we show
how simulations can be a powerful tool in defining and optimizing
the survey strategy of a transiting planet survey. We find the
Pan-Planets project to be competitive with all other existing and
planned transit surveys with the main power being the large 7 square
degree field of view. In the first year we expect to find up to 25
Jupiter-sized planets with periods below 5 days around stars
brighter than V = 16.5 mag. The survey will also be sensitive to
planets with longer periods and planets with smaller radii. After
the second year of the survey, we expect to find up to 9 Warm
Jupiters with periods between 5 and 9 days and 7 Very Hot Saturns
around stars brighter than V = 16.5 mag as well as 9 Very Hot
Neptunes with periods from 1 to 3 days around stars brighter than i'
= 18.0 mag.
Key words: stars: planetary systems - methods: statistical - techniques: photometric
1 Introduction
Thirteen years after the first discovery of a planet revolving around a main-sequence star (Mayor & Queloz 1995) more than 300 extra solar planets are known. The majority of these have been detected by looking for periodic small amplitude variations in the radial velocity of the planet's host star. This method reveals the period, semi-major axis and eccentricity of the planetary orbit, but due to a generally unknown inclination only the minimum mass of the planet can be derived. The situation is different in the case of a planet that transits its host star. The planet blocks a small fraction of the stellar surface resulting in a periodic drop in brightness. In combination with radial velocity measurements the light curve provides many additional parameters like inclination, true mass and radius of the planet.
Transiting planets are subject of many detailed follow-up studies such as measurement of thermal emission using the secondary transit (Knutson et al. 2008; Charbonneau et al. 2008) or measurement of the spin-orbit alignment using the Rossiter-McLaughlin effect (e.g. Winn et al. 2007).
In the past years many transit projects have monitored hundreds of thousands of stars looking for periodic drops in the light curves. A total of about 50 transiting planets are known to date. Remarkably, more than half of the transiting planets have been found in the past year making the transit method equally successful in that period compared to the radial velocity method.
The majority of the recently detected transiting planets have been found by wide-angle surveys targeting bright stars such as WASP, HAT, TrES or XO (McCullough et al. 2005; Noyes et al. 2008; O'Donovan et al. 2007; Pollacco et al. 2006). Also the space mission Corot has contributed by adding four new discoveries (Aigrain et al. 2008). Deep surveys like OGLE targeting highly crowded regions of the Milky Way disk have not been able to keep up with the increased detection rate of all-sky monitoring programs mainly due to limited amount of observation time and a lower number of target stars.
In 2009 Pan-Planets - a new deep transit survey - will start taking first observations. This project will be more powerful than all existing deep surveys because of its by far larger field of view, bigger telescope and faster readout.
With the first detections of transiting extra-solar planets, several groups have started to predict the number of planets that could be found by existing and planned surveys. First estimates were based on optimistic assumptions and have been mostly over-predictions (e.g. Horne 2003). For example the frequency of very close-in planets had been extrapolated from the metallicity biased results of the radial velocity surveys. Late type dwarfs with higher metallicity turned out to have a higher frequencies of close-in planets (Fischer & Valenti 2005) and therefore transit surveys find less planets compared to radial velocity surveys due to a lower average metallicity (Gould et al. 2006). It was further assumed that planets could be found around all stars in the target fields whereas planets transiting giants show a much too faint photometric signal due to the larger radius of the star. In addition, the efficiency of the detection algorithm was not taken into account and all light curves with 2 visible transits were assumed to lead to a detection which is not the case.
More realistic methods have been introduced by
Pepper et al. (2003) and Gould et al. (2006). Both
groups use an analytical approach assuming a stellar and a planetary
distribution and integrating over period, stellar mass, planetary
radius and volume probed taking into account the detection
probability. Similarly, Fressin et al. (2007) modeled the OGLE
survey and compared the predicted distributions to the parameters
actually found by OGLE. In a recent study, Beatty & Gaudi (2008)
generalized the formalism of Gould et al. (2006) in order to
provide a method that can be used to calculate planet yields for any
photometric survey given the survey parameters like number of nights
observed, bandpass, exposure time, telescope aperture, etc. They
applied their method to a number of different planned surveys like
SDSS-II and the Pan-STARSS 3
survey.
In this work we use Monte Carlo simulations to predict the number of planets of the Pan-Planets survey. Our approach is quite general and applicable to any transit survey. Based on stellar and planetary populations we model the survey by constructing realistic light curves and running a detection algorithm on them. In this way we are able to directly include the effects of limb darkening, ingress/egress and observational window functions which have not been included in most previous studies. In addition we introduce a model for correlated noise and study its impact on the efficiency of the detection algorithm.
To optimize the survey strategy of Pan-Planets we want to address the following questions: What is the best observing block size (1 h or 3 h) and how many fields (3 to 7) should we observe? Given the optimized survey strategy, we study how many Very Hot Juptiters (VHJ) and Hot Jupiters (HJ) are expected in the first year and what is the potential of Pan-Planets to find planets with longer periods, such as Warm Jupiters (WJ) or planets with smaller radii, such as Very Hot Saturns (VHS) and Very Hot Neptunes (VHN). We further study whether it will be more efficient to observe the same target fields in the second year of the Pan-Planets survey or to choose new ones.
In Sect. 2 we give a brief overview of the Pan-Planets survey. Section 3 describes in detail the simulations we performed. We present our results in Sect. 4. In order to verify our results we perform a consistency check with the OGLE-III survey by comparing our predicted yield with the actual number of planets found (Sect. 5). Finally we draw our conclusions in Sect. 6.
2 Pan-Planets overview
The Panoramic Survey Telescope and Rapid Response System (PanSTARRS)
is an Air Force funded project aiming at the detection of killer
asteroids that have the potential of hitting the Earth in the near
future. The prototype mission PanSTARRS1 is using a 1.8 m telescope at
the Haleakala Observatories (Maui, Hawaii) to monitor 3
of the
sky over a 3.5 yr period starting in early 2009. The telescope is
equipped with the largest CCD camera in the world to date that samples
a field of 7 sq.deg. on a 1.4 Gigapixel array
(Kaiser 2004) with a pixel-size of 0.258 arcsec.
To make use of the large amount of data that will be collected, a science consortium of institutes from USA, Germany, UK and Taiwan has defined 12 Key Science Projects, out of which one is the Pan-Planets transit survey. A total of 120 h per year have been dedicated to this project during the 3.5 yr lifetime of the survey. The actual observing time will be less due to bad weather and technical downtime. We account for a 33% loss in our simulations.
In the first 2 years, Pan-Planets will observe 3 to 7 fields in the direction of the Galactic plane. Exposure and read-out time will be 30 s and 10s respectively. The observations will be scheduled in 1 h or 3 h blocks. The target magnitude range will be 13.5 to 16.5 mag in the Johnson V-band. The magnitude range is extended to i' = 18 when searching for Very Hot Neptunes (see Sect. 4.6). More detailed informations about Pan-Planets are presented in Afonso et al. (in prep.).
3 Description of the simulations
The goal of this work is to study the expected number of planets that will be detected by the Pan-Planets project as a function of different survey strategies, with a variety of different parameters like number of fields (3 to 7), length of a single observing block (1 h and 3 h) and level of residual red noise (0 mmag, 1 mmag, 2 mmag, 3 mmag and 4 mmag). In total we simulate about 100 different combinations of these parameters for each of 5 different planet populations (see Sect. 3.2).
In our simulations we follow a full Monte-Carlo approach, starting with the simulation of light curves with realistic transit signals. Systematic effects coming from data reduction steps on image basis, such as differential imaging or PSF-photometry are taken into account by adding non-Gaussian correlated noise, the so called red noise (Pont et al. 2006), to our light curves (see Sect. 3.4). We apply a box-fitting-least-squares algorithm to all simulated light curves in order to test whether a transiting planet is detected or not.
For each star in the input stellar distribution (Sect. 3.1) we decide randomly whether it has a planet or not, depending on the fraction of stars having a planet of this type. In the case it has a planet, we randomly pick a planet from the input planet distribution (Sect. 3.2) and create a star-planet pair which is attributed a randomly oriented inclination vector resulting in a transiting or non-transiting orbit (the geometric probability for a transiting orbit depends on stellar radius and semi-major axis of the orbit).
In the case of a transiting orbit, the light curve is simulated based on stellar and planetary parameters and the observational dates we specified (see Sect. 3.5). The shape of the transit is calculated according to the formulae of Mandel & Agol (2002) and includes the effects of ingress/egress and limb-darkening. We add uncorrelated Gaussian (white) and correlated non-Gaussian (red) noise to our light curves. Details about our noise model are given in Sects. 3.3 and 3.4. After the simulation of the light curves, we apply our detection algorithm and our detection cuts as described in Sect. 3.6, and count how many planets we detect.
One simulation run is finished after each star has been picked once. In this way one run represents one possible outcome of the Pan-Planets survey. Since in the majority of cases the star has no planet or the inclination is such that no transits are visible, there are in general only a few transiting light curves per run. For each planet population and each set of survey parameters we simulate 25 000 runs. For the selected survey strategy we increase the precision to 100 000 runs. The numbers we list in our results are averages over these runs. The scatter of the individual outcomes allows us to derive errors for our estimates.
3.1 Input stellar distribution
We make use of a Besançon
model
(Robin et al. 2003) for the spectral type and brightness
distributions of stars in our target fields. A model of 1 sq.deg
centered around
,
(l
= 54.5, b = -4.2) is scaled to the actual survey area assuming a
constant density. The parameters taken from the model are stellar mass
,
effective temperature
,
surface gravity
,
metallicity
and apparent
MegaCam
i'-band AB-magnitude mi'. The model also provides colors which we
use to determine the apparent Johnson V-band magnitude mV,
according to the following formula derived by
Smith et al. (2002):
![]() |
(1) |
The stellar radii




![$\rm [Fe/H]$](/articles/aa/full_html/2009/05/aa10754-08/img13.gif)
In total we find 3 440 F, G, K and M dwarfs per sq.deg. that are not
saturated (i.e.
mag) and are brighter than our radial
velocity follow-up limit (i.e.
mag).
Figure 1 shows the input stellar distribution.
For VHN we extend the target magnitude range to
mag. We find 34 000 M dwarfs in this range.
![]() |
Figure 1:
Total number and magnitude histogram for F, G, K and M dwarfs ( top to bottom) with
|
Open with DEXTER |
3.2 Input planet distributions
We test five different planetary populations:
- 1.
- very Hot Jupiters (VHJ), with radii of 1.0-1.25 RJ and periods between 1 and 3 days;
- 2.
- hot Jupiters (HJ), with radii of 1.0-1.25 RJ and periods between 3 and 5 days;
- 3.
- warm Jupiters (WJ), with radii of 1.0-1.25 RJ and periods between 5 and 10 days;
- 4.
- very Hot Saturns (VHS), with radii of 0.6-0.8 RJ and periods between 1 and 3 days;
- 5.
- very Hot Neptunes (VHN), with radii of 0.3 RJ and periods between 1 and 3 days.
Our predicted yields depend on the frequency of stars that have a
planet for each of the five population. These frequencies are not
known to a very good precision and not many estimates have been
published so far. Gould et al. (2006) performed a detailed
study of the OGLE-III survey and derived frequencies of Very Hot
Jupiters and Hot Jupiters by comparing the number of detected planets
in the OGLE-III survey to the number of stars the survey was sensitive
to. They found at 90% confidence level
of all late type dwarfs to have a VHJ and
to have an HJ.
Fressin et al. (2007) published comparable results analyzing
the same survey.
For VHJ and HJ we use the frequencies published by Gould et al. (2006). The frequency of WJ we speculate to be the same as for HJ which is consistent with the OGLE-III results (see Sect. 5). Further we assume the frequencies for VHS and VHN to be 0.714% (same as for VHJ) and 5% respectively.
3.3 White noise model
For the white noise in our light curves we add four different Gaussian components: stellar photon noise, sky background, readout and scintillation noise. The photon noise of each star is estimated using a preliminary exposure time calculator which has been calibrated by observations taken during a pre-commissioning phase of the PanSTARRS1 telescope. We assume the sky background to be 20.15 mag per square arcsecond which corresponds to a seven day distance to full moon. The readout noise is assumed to be 8 e- per pixel. The scintillation noise is estimated to be 0.5 mmag according to the formula of Young (1967,1993) and is only of importance at the very bright end of our target distribution. Figure 2 shows the white noise as function of magnitude as well as the individual contributions.
For our calculations we assume a seeing of 1.2 arcsec, airmass of 1.4, extinction coefficient of 0.08 and PSF fitting radius of 1.0 arcsec. At the faint end (i' = 18 mag) the number of photons is on the order of 15 500 for the object and 7800 for the sky and therefore well outside the Poisson statistics regime.
![]() |
Figure 2: Total white noise and contributions of the four components. |
Open with DEXTER |
3.4 Residual red noise model
Table 1:
Dimensionless relative amplitudes Ai' and
timescales
of our different red noise models. Models 1 to 3 are fixed parameter models whereas for the others we
draw random values within a given range for each single
light curve individually. The timescales are chosen to cover
the range of expected transit durations.

with


Application of algorithms to remove systematic effects, such as Sysrem (Tamuz et al. 2005) or TFA (Kovács et al. 2005) have been successfully applied by several groups resulting in a significant reduction of the level of red noise (e.g. Snellen et al. 2007). However, a small fraction of the correlated noise always remained.
In our simulations we want to account for this residual red noise (RRN). A simple model would be to increase the level of Gaussian noise by a certain amount and therefore assume that the correlated nature is of minor importance. For studies based only on S/N calculations one could also use a parameterization like Eq. (2). Since we are simulating light curves, we want to introduce a different approach. We model the RRN by adding superimposed sine waves of different wavelengths and amplitudes. This allows us to include the effects the correlated noise has on the efficiency of the detection algorithm, which could get confused by noise that is correlated on timescales of a typical transit duration.
We add RRN according to the following model:
![]() |
(3) |
with normalized amplitude Ai, timescale

For each model we start with relative amplitudes Ai' which are
normalized in such a way that the rms of the added RRN
(
)
is of value 1 mmag, 2 mmag, 3 mmag or 4 mmag:
![]() |
(4) |
In order to analyze the influence of the timescales and amplitudes on our results we construct a total of 9 different red noise models with each of them having 3 or 4 components. Table 1 gives an overview of the parameters of our red noise models. We refer to models 1 to 3 as ``fixed parameter'' models because for these we select arbitrary fixed values for Ai' and


![]() |
Figure 3:
|
Open with DEXTER |
3.5 Epochs of the observations
For each year the Pan-Planets survey has been granted a total of 120 h hours which will be executed in 1 h or 3 h blocks. Assuming a 33% loss due to bad weather we expect the data to be taken in 81 or 27 nights per year depending on our survey strategy. The actual epochs of the observations we use to construct our light curves are computed in the following way. For each night in which our target field is visible for at least 3 h we calculate the range of visibility, namely the time the field is higher than airmass 2 on the sky. This results in a 183 day period starting on April 26th and ending on October 24th. We randomly pick nights during the period of visibility and place the observing block arbitrarily within the time span our target is higher than airmass 2, as calculated earlier.
In our simulations we test 5 different scenarios with alternate observations of 3 to 7 fields during the observing block. The time for one exposure and readout is assumed to be 40 s. Therefore the different number of fields transform into cycle rates between 120 s and 280 s. Using the selected nights, the random position of a block within a night and the cycle rate we construct a table of observational dates which we use as input to the light curve simulations. For each simulation run (which represents one possible outcome of the survey) we draw new observational dates. Table 2 summarizes the observational parameter depending on the survey strategy.
Table 2: Cycle rate and number of data points per night and year depending on observational strategy.
3.6 Light curve analysis
Each simulated light curve is analyzed by our detection algorithm
which is a box-fitting-least-squares (BLS) algorithm proposed by
Kovács et al. (2002). The program folds the light curves with
trial periods in the range from 0.9 to 9.1 days and finds the best
fitting box corresponding to a fractional transit
length
between 0.01 and 0.1. For each detection the BLS algorithm provides period, S/N and the
number of individual transits. For a successful detection we require
the period found to match the simulated period within 0.2% (see
Fig. 4). In addition, we impose the S/N to be
larger than 16 (see Sect. 3.7) and the number of transits
to be at least equal to 3. The planet is also considered being
detected if the measured period is half or twice the simulated period
(to within 0.2%). This can easily happen in case of unevenly sampled
light curves. For later analysis we store all input parameters of the
simulation and output parameters of the detection algorithm in a
table.
![]() |
Figure 4:
Deviation of the detected period
|
Open with DEXTER |
3.7 Signal-to-noise cut
To model a transit survey, it is very important to have a transparent
and reproducible procedure of applying cuts in the process of
selecting the candidates. The most important value is the minimum
S/N. The S/N of a transit light curve is defined as the transit depth
divided by the standard deviation of the photometric average of all
measurements taken during a transit. For a light curve with Nuniformly spaced data points with individual Gaussian error ,
transit depth
and a fractional transit length
this
is:
![]() |
(5) |
In the presence of red noise the value of the S/N is reduced. Also the actual shape of the transit, which is determined by limb-darkening and ingress/egress, has an impact on the S/N. This effect is included implicitly in our simulation.
Since the probability of finding a planet is small, the majority of
transit surveys use a low S/N cut of about 10. This results in a high
number of statistical and physical false
positives and has
made it necessary to include non-reproducible selection procedures
such as ''by-eye'' rejection. Pushing the S/N cut to the detection
limit makes it therefore difficult to model the detection
efficiency.
In the Pan-Planets survey we expect to find a very high number of
candidates already in the first year which will require a high amount
of radial velocity follow-up resources. The best candidates have the
highest S/N and will be followed-up first. We will most likely not be
able to follow-up all candidates down to the detection threshold of 12 and therefore use a somewhat larger S/N cut. In this work we
calculate the expected number of detections using an S/N cut of 16.
4 Results of the Monte Carlo simulations
In this section we summarize the results of a total of 7.6 million simulation runs. The computation time was 230 000 CPU hours which we distributed over a 486 CPU beowulf cluster.
In Sect. 4.1 we show which block size (1 h or 3 h) is more efficient for the Pan-Planets survey. In Sect. 4.2 we compare the different RRN models. Section 4.3 addresses the question of the optimal number of fields (3 to 7). In Sect. 4.4 we summarize the actual number of VHJ, HJ, WJ, VHS we expect to find using our preferred survey strategy. Finally, we show the results in the case of observing the same fields during the second year of the survey instead of monitoring new ones (Sect. 4.5). In Sect. 4.6 we study the potential to find Very Hot Neptunes transiting M dwarfs.
Error estimates are only given for the final numbers in
Sects. 4.4 and 4.5. All numbers
we present are scaled from the 1 sq.deg. Besançon model to the
actual survey area of
sq.deg assuming a constant
spectral type and magnitude distribution and a homogeneous density.
In order to check whether there are 7 fields of comparable density, we
count the total number of stars in the USNO-A2.0 catalog and compare
it to the total number of stars in the Besançon model for a set of
different Galactic longitudes (Table 3). We assumed
an average color (
)
of 0.25 mag. In the range
the number of stars in the Besançon model
agrees well with the number of stars. The USNO density varies at a
level of 30% with the average being
14 000, close to the
density we assume in our simulations (l = 54.5). With a diameter of 3 deg. a total of 7 Pan-Starrs fields fit in this range.
Table 3: Total number of stars per sq. deg. according to the USNO-A2.0 catalog and the Besançon model.
4.1 Influence of the size of the observing blocks
We investigate the influence of the observing block size on the number of detections in the Pan-Planets survey. Table 4 lists the average number of VHJ and HJ found with 1 h and 3 h blocks after the application of our detection cuts, as described in Sect. 3.7.
The first three columns list the planet population and the survey strategy (i.e. number of fields and observing block size). The fourth column shows the average numbers of all simulated transiting planet light curves having an S/N of 16 or more (without requiring 3 transits and without running the detection algorithm). Here the numbers are very similar comparing the 1 h to the 3 h block strategies.
To understand this, one has to consider that a planet spends a certain
fraction of its orbit in transit phase (also known as fractional
transit length ). This fraction depends mainly on the
inclination and period of the orbit as well as the radius of the host
star. For a given
the average number of points in transit (
)
only depends on the total number of observations Nand is therefore independent of the block size. The same applies to
the S/N which, for fixed transit depth and photometric noise
properties, depends only on the number of points in transit to a good
approximation. Therefore, if only a minimum S/N is required, the
number of detections is comparable for a strategy with 1 h blocks and
with 3 h blocks, with minor differences arising from limb-darkening and
ingress/egress effects.
Although the number of points in transit is the same for both strategies, one 3 h block covers on average a bigger part of the transit compared to a 1 h block. As a consequence the average number of individual transits must be lower in the case of 3 h blocks. If we impose the additional cut of requiring at least 3 transits to be visible in the light curve (Col. 5), the expected number of planets found is lower for the 3 h blocks compared to the 1 h blocks. With a 3 h block strategy the number of light curves passing the S/N cut and having 3 or more transits is on average 53% lower for HJ and 26% lower for VHJ. For the longer period HJ this effect is stronger due to the fact that the number of visible transits is lower in general.
In order the planet to be considered detected (as described in
Sect. 3), we not only require
and at least 3 visible transits, but also that the BLS detection algorithm finds the
correct period (allowing for twice and half the correct value). The
impact of this additional selection cut is shown in Cols. 6-10 for
light curves with 0 mmag, 1 mmag, 2 mmag, 3 mmag and 4 mmag
RRN
. Without RRN, most
planets are found by the BLS algorithm. The loss is marginally higher
in the case of 3 h blocks which is a consequence of the generally lower
number of transits, since the BLS algorithm is more efficient if more
transits are present. Comparing the results for 1 h and 3 h blocks we
find that in case the RRN level is 2 mmag, the number of detected
planets without RRN is on average 59% lower for HJ and 30% lower for
VHJ in the 3 h block case.
Including RRN, fewer planets are detected by the BLS algorithm and the discrepancy between 1 h and 3 h blocks increases. For a typical RRN level of 2 mmag we find on average 71% less HJ and 45% less VHJ with 3 h blocks compared to 1 h blocks.
As an additional test we perform the same analysis for a campaign with twice the amount of observing time spread over 2 years. This would correspond to a strategy where we stay on the same target fields in the second year of the Pan-Planets survey. Also in this case, 1 h blocks are more efficient than 3 h blocks. Assuming the RRN level is 2 mmag, we find that the number of detected planets is on average 34% lower for HJ and 59% lower for VHJ in the 3 h block case. The details of the simulations for a 2 yr campaign can be found in Sect. 4.5. In the following we restrict our results to 1 h blocks.
Table 4: Influence of the block size shown on the basis of the number of planets detected in a 1 yr campaign after applying different detection cuts.
Table 5: Influence of the red noise model for 1 h blocks and a RRN level of 2 mmag.
4.2 Influence of the residual red noise model
In this section we compare the results of nine different RRN models which have been introduced in Sect. 3.4. In addition, we compare the red noise models to a scenario where we add additional uncorrelated white noise by the same amount as the RRN level. Table 5 shows the number of HJ and VHJ found with a 1 h block strategy and 2 mmag RRN for each of the 9 different red noise models, as well as for the increased white noise model.
In general the increased white noise model results in a significantly higher number of detections compared to the RRN models (on average 22% and 39% higher for VHJ and HJ respectively). This shows that the effect of the RRN on the efficiency of the BLS algorithm is strong and needs to be taken into account in our simulations.
Comparing the individual RRN models to each other we find that for the fixed parameter models (1 to 3) the number of detections is 8% and 13% higher for VHJ and HJ respectively than for the random models (4 to 9). The individual results of the random models are all very similar and vary only by a few percent. In the following we restrict our results to the red noise model 4, since it is the most general of all models with 4 components and random timescales ranging from 0 to 400 min.
4.3 Influence of the number of fields
In order to optimize the survey with respect to the number of
alternating fields monitored during an observing block, we compare the
number of detections for each of the 5 strategies (3 to 7 fields). We
do not test more than 7 fields, because it is not sure if we can find
a higher number of fields with comparable density (see
Sect. 4). Note also, that with more than 7 fields, the number
of data points per light curve would be less than 1000 and the cycle
rate longer than 5 min which would complicate the process of
eliminating false positives on the basis of the light curve
shape. We limit our simulations to the above selected 1 h blocks
(see Sect. 4.1) and 2 mmag RRN of model 4 (see
Sect. 4.2).
In general, the total number of detections depends on the number of fields in two counteracting ways: on the one hand, observing more fields results in more target stars and therefore more transiting planet systems that can be detected; on the other hand, observing more fields results in a lower number of data points per light curve and thus the S/N of each transit candidate is shifted to a lower value. The latter effect is stronger for faint stars because the S/N is generally lower whereas for brighter stars the S/N is high enough in most cases.
Table 6: Number of planets found in the first year depending on the number of target fields, assuming 1 h blocks and 2 mmag RRN (model 4).
The number of detections in the first year of the Pan-Planets survey for different number of fields is shown in Table 6. For all planet populations (i.e. VHJ, HJ, WJ and VHS) we find more planets with a higher number of fields. The loss in S/N is over-compensated by the higher number of target stars. In Fig. 5 we show the S/N distributions of VHJ detections for a 3 field and a 7 field strategy. The S/N distribution of VHJ peaks at a higher level than our cut of 16, even for the 7 field strategy, which explains why observing more fields results in more detections.
![]() |
Figure 5: Normalized S/N distribution of VHJ detections for a 3 field strategy (black) and a 7 field strategy (red) assuming 1 h blocks and 2 mmag RRN (model 4). |
Open with DEXTER |
In the case of a 2 yr campaign the situation is the same. For all four planet populations it is more efficient to observe a higher number of fields (see Table 7). Therefore we conclude that observing 7 fields is the most efficient strategy and restrict our results in the following to 7 fields.
Table 7: Number of planets found in 2 yr depending on the number of target fields, assuming 1 h blocks and 2 mmag RRN (model 4).
4.4 The expected number of planets in the Pan-Planets survey
In the previous sections we have identified our preferred survey strategy with 1 h blocks and alternating among 7 fields. In addition we selected RRN model 4 as our preferred one. For these parameters we performed more detailed simulations in order to calculate the expected number of detections of the Pan-Planets project (including error estimates) and to study the parameter distributions of the detected planets in detail. For each of 4 different RRN levels (1 mmag, 2 mmag, 3 mmag and 4 mmag) and 4 planet populations (VHJ, HJ, WJ and VHS) we performed 25 000 simulation runs. The number of detections depending on the level of RRN are shown in Table 8. Figure 6 shows the cumulative distribution of the host star brightness for each planet populations for 2 mmag of RRN (model 4).
Our predicted numbers are affected by two sources of uncertainties. The first and dominant one is the uncertainty of the planet frequency taken from Gould et al. (2006) which is caused by the low number statistics of the OGLE detections. This uncertainty is not included in Table 8 and must be taken into account by scaling all of our HJ results by a factor of 1+1.37-0.59 and all of our VHJ results by a factor of 1+1.10-0.54, as published in Gould et al. (2006).
The second uncertainty is a direct result of our simulations. Each simulation run represents one possible outcome of 1 sq.deg. of the Pan-Planets survey. Since the simulated observational epochs change from one run to the other and since in each run different stars are attributed to planets with different orbital parameters, an intrinsic scatter in the number of planets in found in each run. The combination of 49 randomly chosen runs represents one possible outcome of the full 49 sq.deg. survey (7 fields). From the histogram of these combinations we derive 68% confidence intervals for our predicted numbers.
Table 8: Number of planets found in the first year for our selected survey strategy as a function of residual red noise level (model 4).
![]() |
Figure 6: Cumulative host star brightness distributions for each planet population for a 1 yr campaign with 2 mmag RRN (model 4). |
Open with DEXTER |
![]() |
Figure 7: Cumulative host star brightness distributions for each planet population for a 2 yr campaign with 2 mmag RRN (model 4). |
Open with DEXTER |
4.5 Number of expected planets in a two year campaign
The Pan-Planets project has a lifetime of 3.5 years. In the previous sections we focus mainly on the first year of the survey. In this section we show the results of our simulations for a 2 year campaign. In particular, we address the question whether the project is more efficient if we stay on the same fields or if we choose new targets assuming that we find fields with similar densities (see Sect. 6).
Table 9 shows the number of planets detected in a 2 yr campaign for four different levels of RRN. Except for VHJ, we more than double the number of detections for each planet population. For the longer period WJ the gain is a factor of 3. In a 1 yr campaign most of these planets show less than 3 transits and are not detected. Adding the observations of the second year, the number of transits increases and many of the previously undetected planets are found.
In addition, staying on the same fields in the second year increases the S/N of all transit light curves, due to the higher number of data points taken during a transit. Planets that have an insufficiently high S/N after the first year are detected after the second year. This is particularly true for VHS.
Figure 8 shows the fraction of all transiting Jupiter-sized planets (VHJ, HJ, WJ) that are detected in the first year of the survey (lower black line) and in the 2 yr campaign (upper red line). In the first year, the average efficiencies are 26.3%, 10.6% and 4.3% for VHJ, HJ and WJ respectively. Planets that have been missed do not have the required S/N, show less than 3 transits in the light curve, or the BLS algorithm found a wrong period. Extending the survey to the second year increases the efficiency significantly to 39.8%, 24.0% and 14.3% for VHJ, HJ and WJ respectively.
![]() |
Figure 8: Fraction of all transiting VHJ, HJ and WJ that are detected as a function of period for a 1 yr campaign ( lower black line) and a 2 yr campaign ( upper red line). |
Open with DEXTER |
Table 9: Number of planets found with 160 h of observations in 2 yrs for our selected survey strategy as a function of residual red noise level (model 4).
4.6 The detection of Very Hot Neptunes
In this section we study the potential to find Very Hot Neptunes transiting M dwarfs. The radius ratio between planet and star is much higher for low mass stars which results in much deeper transits and therefore a higher detection probability. According to the Besançon model there are a total of 34 000 M dwarfs brighter than AB-magnitude mi' = 18 mag in 7 fields of 7 sq.deg. each. These objects are particularly interesting, since the composition of planets in this mass range is rather unknown (gaseous, icy or rocky). Also the habitable zone is much closer to the star due to its lower surface temperature. Note, that only one planet transiting an M dwarf has been detected so far.
We consider all transiting VHN candidates down to host star brightnesses of mi' = 18 mag to be interesting objects, although the spectroscopical follow-up will be very challenging. New high resolution near infrared spectrographs will help to confirm these very red objects.
To study the potential of Pan-Planets to find transiting VHN we perform simulations for the whole input stellar distribution and analyze the spectral type distribution of the hosts stars of all successful detections (Fig. 9). The Pan-Planets survey is sensitive to close-in Neptune-sized planets around late K and early M dwarfs if the frequency of these stars hosting Neptunes is as large as 5%. The number of VHN detections after the first and the second year is listed in Table 10 for 4 different residual rednoise levels. Assuming 2 mmag of RRN we expect to find 3 VHN after the first and 9 VHN after the second year.
Further we analyze the distance distributions of all detected VHJ, VHS and VHN systems (Fig. 10). The volume probed strongly depends on the radius of the planet. For lower mass radius the transit depth is generally smaller and therefore the photometric precision needed to detect the transits must be higher, which is only the case for closer and thus brighter systems. Note that for HJ and WJ the distance distributions are very similar to the VHJ distribution.
![]() |
Figure 9: Host star spectral type distribution of all detected VHN for a 2 yr campaign with 2 mmag RRN (model 4). Pan-Planets is sensitive to VHN transiting late K and early M dwarfs. |
Open with DEXTER |
Table 10: Number of VHN detections after the first and second year of the Pan-Planets survey for our selected survey strategy for different residual red noise levels (model 4).
![]() |
Figure 10: Distance distribution of the detected VHJ, VHS and VHN for a 2 yr campaign with 2 mmag RRN (model 4). Due to the lower transit depth smaller planets can only be detected around closer stars. |
Open with DEXTER |
5 Consistency check with the OGLE-III survey
Gould et al. (2006) have modeled the OGLE-III survey in order to derive absolute frequencies of HJ and VHJ. In our simulations we are using these frequencies to predict the number of detections for the Pan-Planets survey. In order to verify our results we performed a consistency check by modeling the OGLE-III survey and comparing the results to the actual number of planets found. We limit this test to the 3 Carina fields where 3 planets have been found (1 HJ, 2 VHJ). The 2 bulge fields that have also been observed during the OGLE-III campaign are more difficult to model due to a stronger blending and a higher uncertainty in the input stellar distribution.
We obtained the Besançon model population of the 3 Carina fields
(CAR100, CAR104, CAR105) for stars in the magnitude range of
.
The overall noise level as a function of magnitude
has been determined by Gould et al. (2006) to be:
![]() |
(6) |
In their simulations Gould et al. (2006) did not include correlated noise sources, instead they account for systematics by using an increased S/N cut. In order to be as consistent as possible we follow the same procedure and do not split the over-all noise in red and white noise components (as we did in the Pan-Planets simulations). We use the radius and period distributions for HJ and VHJ introduced in Sect. 3.2. The epochs of the observations were taken from the light curve of OGLE-TR-74
![[*]](/icons/foot_motif.gif)
After simulating the light curves in the same way as described in
Sect. 3 we run the BLS algorithm and check for a correct
period recovery. In addition we apply the following cuts which have
been used by the OGLE group and are summarized in detail in Sect. 3
and Table 1 in Gould et al. (2006): the transit depth must be smaller than 0.04 mag (
3.62%); the S/N greater than 11.6; the signal detection efficiency
larger than 3.8;
the number of transits is required to be at least 3; and finally, the
color (V-I)0 must be greater than 0.4. Note that we have not
imposed any cut on the transit depth in our simulations for the
Pan-Planets survey since a Jupiter-sized planet transiting an M dwarf
can have a fairly high transit depth. Further we do not use a color
since in our simulations we include only late type dwarfs a priori.
In total we simulated 50 000 runs for each of the five planet populations. On average we find 2.18 VHJ and 1.46 HJ which is in reasonable good agreement with the actual number of 2 VHJ and 1 HJ found by the OGLE group.
According to our simulations the OGLE-III carina survey was not sensitive to one of the other 3 planet populations we tested. We find on average 0.45 WJ, 0.12 VHS and zero VHN which is agreement with none being found by OGLE.
6 Conclusion
The aim of this work was to study the influence of the survey strategy on the efficiency of the Pan-Planets project and to predict the number of detections for an optimized strategy.
Our calculations are based on the simulation of realistic light curves including the effects of limb-darkening, ingress/egress and observational window functions. In addition we have introduced a model to simulate correlated (red) noise which allows us to include the effects of correlated noise on the efficiency of the BLS detection algorithm. Our approach can be applied to any transit survey as well.
Below we summarize the caveats and assumptions that were made in our simulations:
- Our results depend on the spectral type and magnitude distribution of the Besançon model. The model does not include second order substructure such as spiral arms.
- We neglect the effects of blending. Due to crowding into the direction of the Galactic disk some stars are blended by neighboring sources.
- We assume all planets that are detected by the BLS-algorithm to be followed-up and confirmed spectroscopically. In particular, we assume that no true candidate is rejected by any candidate selection process. The detailed follow-up strategy of the Pan-Planets survey will be presented in Afonso et al. (in prep.).
- Our simulations are done for 1 sq.deg. and the results are scaled to the actual survey area. We assume that all fields (3, 4, 5, 6 or 7 case) have homogeneous densities and non-varying (or similar) stellar populations. Simple number counts on the USNO-catalog showed that we can find up to 7 fields with similar total number of stars (see Sect. 4). For a larger number of fields, the assumption of a constant density might be too optimistic since we are restricted to fields that are close to each other in order to keep the observational overhead low.
- Our results directly scale with the assumed planet frequencies. The values of 0.14% and 0.31% we use for VHJ and HJ have uncertainties of a factor of 2. For WJ, VHS and VHN we have used hypothetical values of 0.31%, 0.14% and 5% respectively. After completion of the Pan-Planets survey we will be able to derive more accurate absolute frequencies for all five planet populations.
- The quality of the data is assumed to be homogeneously good over the whole detector area. Bad pixel regions and gaps between the individual CCDs are not taken into account and result in an effective field of view that is smaller than 7 sq.deg.
For an RRN level of 2 mmag we expect to find up to 15 VHJ and 10 HJ in the first year around stars brighter than V = 16.5 mag. The survey will also be sensitive to planets with longer periods (WJ) and smaller radii (VHS and VHN). Assuming that the frequencies of stars with WJ and VHS is 0.31% and 0.14% respectively, we expect to find up to 2 WJ and 3 VHS in the same magnitude range.
We found that observing the same fields in the second year of the 3.5 yr lifetime of the survey is more efficient than choosing new fields. We expect to find up to 24 VHJ, 23 HJ, 9 WJ and 7 VHS. In particular for longer periods (HJ and WJ) and smaller radii (VHS) we will more than double the number of detections of the first year if we continue to observe the same targets.
We have investigated the potential of the Pan-Planets survey to detect VHN transiting M dwarfs brighter than i' = 18 mag. Assuming the frequency of these objects is 5%, we expect to find up to 3 detections in the first year and up to 9 detections observing the same fields in the second year.
As a consistency check we modeled the OGLE-III Carina survey and found 2.18 VHJ, 1.46 HJ, 0.45 WJ, 0.12 VHS and zero VHN which is in agreement with the 2 VHJ and 1 HJ and the zero WJ, VHS and VHN that have been actually detected.
Acknowledgements
We thank the referee Scott Gaudi for the constructive feedback. His comments and suggestions have helped us to identify the optimal survey strategy of the Pan-Planets project as well as improving the presentation.
References
- Aigrain, S., Barge, P., Deleuil, M., et al. 2008, in 14th Cambridge Workshop on Cool Stars, Stellar Systems, and the Sun, ed. S. P. P. S. D. E. M. B. J. Messina, ASP Conf. Ser., 384, 70 (In the text)
- Beatty, T. G., & Gaudi, B. S. 2008, ArXiv e-prints, 804 (In the text)
- Charbonneau, D., Knutson, H. A., Barman, T., et al. 2008, ArXiv e-prints, 802
- Claret, A. 2000, A&A, 363, 1081 [NASA ADS] (In the text)
- Claret, A. 2004, A&A, 428, 1001 [NASA ADS] [CrossRef] [EDP Sciences] (In the text)
- Fischer, D. A., & Valenti, J. 2005, ApJ, 622, 1102 [NASA ADS] [CrossRef] (In the text)
- Fressin, F., Guillot, T., Morello, V., & Pont, F. 2007, A&A, 475, 729 [NASA ADS] [CrossRef] [EDP Sciences] (In the text)
- Gould, A., Dorsher, S., Gaudi, B. S., & Udalski, A. 2006, Acta Astron., 56, 1 [NASA ADS] (In the text)
- Horne, K. 2003, in Scientific Frontiers in Research on Extrasolar Planets, ed. D. Deming, & S. Seager, ASP Conf. Ser., 294, 361 (In the text)
- Kaiser, N. 2004, in Ground-based Telescopes, ed. J. M. Oschmann, Jr., Proc. SPIE, 5489, 11 (In the text)
- Knutson, H. A., Charbonneau, D., Allen, L. E., Burrows, A., & Megeath, S. T. 2008, ApJ, 673, 526 [NASA ADS] [CrossRef]
- Kovács, G., Zucker, S., & Mazeh, T. 2002, A&A, 391, 369 [NASA ADS] [CrossRef] [EDP Sciences] (In the text)
- Kovács, G., Bakos, G., & Noyes, R. W. 2005, MNRAS, 356, 557 [NASA ADS] [CrossRef] (In the text)
- Mandel, K., & Agol, E. 2002, ApJ, 580, L171 [NASA ADS] [CrossRef] (In the text)
- Mayor, M., & Queloz, D. 1995, Nature, 378, 355 [NASA ADS] [CrossRef] (In the text)
- McCullough, P. R., Stys, J. E., Valenti, J. A., et al. 2005, PASP, 117, 783 [NASA ADS] [CrossRef]
- Noyes, R. W., Bakos, G. Á., Torres, G., et al. 2008, ApJ, 673, L79 [NASA ADS] [CrossRef]
- O'Donovan, F. T., Charbonneau, D., Bakos, G. Á., et al. 2007, ApJ, 663, L37 [NASA ADS] [CrossRef]
- Pepper, J., Gould, A., & Depoy, D. L. 2003, Acta Astron., 53, 213 [NASA ADS] (In the text)
- Pollacco, D. L., Skillen, I., Cameron, A. C., et al. 2006, PASP, 118, 1407 [NASA ADS] [CrossRef]
- Pont, F., Zucker, S., & Queloz, D. 2006, MNRAS, 373, 231 [NASA ADS] [CrossRef] (In the text)
- Robin, A. C., Reylé, C., Derrière, S., & Picaud, S. 2003, A&A, 409, 523 [NASA ADS] [CrossRef] [EDP Sciences] (In the text)
- Smith, J. A., Tucker, D. L., Kent, S., et al. 2002, AJ, 123, 2121 [NASA ADS] [CrossRef] (In the text)
- Snellen, I. A. G., van der Burg, R. F. J., de Hoon, M. D. J., & Vuijsje, F. N. 2007, A&A, 476, 1357 [NASA ADS] [CrossRef] [EDP Sciences] (In the text)
- Tamuz, O., Mazeh, T., & Zucker, S. 2005, MNRAS, 356, 1466 [NASA ADS] [CrossRef] (In the text)
- Winn, J. N., Johnson, J. A., Peek, K. M. G., et al. 2007, ApJ, 665, L167 [NASA ADS] [CrossRef] (In the text)
- Young, A. T. 1967, AJ, 72, 747 [NASA ADS] [CrossRef]
- Young, A. T. 1993, Observatory, 113, 41 [NASA ADS]
Footnotes
- ...
model
- http://bison.obs-besancon.fr/modele/
- ...
MegaCam
- http://www.cfht.hawaii.edu/Instruments/Imaging/Megacam/
- ... dwarfs
- We refer to dwarfs as stars of luminosity class IV-VI.
- ...
length
- The fractional transit length is defined as the transit duration divided by the period.
- ...
positives
- Statistical false positives are purely noise generated detections whereas physical false positives are true low amplitude variations (like e.g. in a blended binary system).
- ...
RRN
- We restrict ourself here to RRN of model 4 - our favored model (see Sect. 4.2).
- ...
shape
- It is important to well sample the ingress/egress part of the transits which has a duration of approximately 15-20 min.
- ...
OGLE-TR-74
- One of the OGLE-III candidates.
- ... efficiency
- Quality parameter provided by the BLS algorithm for each detection.
All Tables
Table 1:
Dimensionless relative amplitudes Ai' and
timescales
of our different red noise models. Models 1 to 3 are fixed parameter models whereas for the others we
draw random values within a given range for each single
light curve individually. The timescales are chosen to cover
the range of expected transit durations.
Table 2: Cycle rate and number of data points per night and year depending on observational strategy.
Table 3: Total number of stars per sq. deg. according to the USNO-A2.0 catalog and the Besançon model.
Table 4: Influence of the block size shown on the basis of the number of planets detected in a 1 yr campaign after applying different detection cuts.
Table 5: Influence of the red noise model for 1 h blocks and a RRN level of 2 mmag.
Table 6: Number of planets found in the first year depending on the number of target fields, assuming 1 h blocks and 2 mmag RRN (model 4).
Table 7: Number of planets found in 2 yr depending on the number of target fields, assuming 1 h blocks and 2 mmag RRN (model 4).
Table 8: Number of planets found in the first year for our selected survey strategy as a function of residual red noise level (model 4).
Table 9: Number of planets found with 160 h of observations in 2 yrs for our selected survey strategy as a function of residual red noise level (model 4).
Table 10: Number of VHN detections after the first and second year of the Pan-Planets survey for our selected survey strategy for different residual red noise levels (model 4).
All Figures
![]() |
Figure 1:
Total number and magnitude histogram for F, G, K and M dwarfs ( top to bottom) with
|
Open with DEXTER | |
In the text |
![]() |
Figure 2: Total white noise and contributions of the four components. |
Open with DEXTER | |
In the text |
![]() |
Figure 3:
|
Open with DEXTER | |
In the text |
![]() |
Figure 4:
Deviation of the detected period
|
Open with DEXTER | |
In the text |
![]() |
Figure 5: Normalized S/N distribution of VHJ detections for a 3 field strategy (black) and a 7 field strategy (red) assuming 1 h blocks and 2 mmag RRN (model 4). |
Open with DEXTER | |
In the text |
![]() |
Figure 6: Cumulative host star brightness distributions for each planet population for a 1 yr campaign with 2 mmag RRN (model 4). |
Open with DEXTER | |
In the text |
![]() |
Figure 7: Cumulative host star brightness distributions for each planet population for a 2 yr campaign with 2 mmag RRN (model 4). |
Open with DEXTER | |
In the text |
![]() |
Figure 8: Fraction of all transiting VHJ, HJ and WJ that are detected as a function of period for a 1 yr campaign ( lower black line) and a 2 yr campaign ( upper red line). |
Open with DEXTER | |
In the text |
![]() |
Figure 9: Host star spectral type distribution of all detected VHN for a 2 yr campaign with 2 mmag RRN (model 4). Pan-Planets is sensitive to VHN transiting late K and early M dwarfs. |
Open with DEXTER | |
In the text |
![]() |
Figure 10: Distance distribution of the detected VHJ, VHS and VHN for a 2 yr campaign with 2 mmag RRN (model 4). Due to the lower transit depth smaller planets can only be detected around closer stars. |
Open with DEXTER | |
In the text |
Copyright ESO 2009
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.