Issue |
A&A
Volume 518, July-August 2010
Herschel: the first science highlights
|
|
---|---|---|
Article Number | L133 | |
Number of page(s) | 7 | |
Section | Letters | |
DOI | https://doi.org/10.1051/0004-6361/201014626 | |
Published online | 16 July 2010 |
Online Material
![]() |
Figure 2:
The 250 |
Open with DEXTER |
![]() |
Figure 6:
The 250, 350 and 500 |
Open with DEXTER |
Table 1: Observation log.
Appendix A: Data reduction
The PACS data were processed in the Herschel interactive
analysis
environment HIPE (v3.0), applying the standard pipeline steps. The
flux conversion was done using version 5 of the response
calibration.
Signal glitches due to cosmic ray impacts were masked out in two steps.
First the PACS photMMTDeglitching task in HIPE was applied on the
detector
timeline. Then a first coarse map was projected, which is then used
as a reference for the second level deglitching HIPE task
IIndLevelDeglitch. In the detector time series we masked the region
around the source
prior to applying a high-pass filter to remove the low frequency
drifts.
The scale of the high pass filter was taken to be half the length of an
individual scan leg on the sky, i.e. 3.7.
The detector time series signals were then summed up into a map using
the PACS photProject task. The pixel scale for the 70 and 100
m
maps was set to 1
,
while the scale for the 160
m map was 2
.
For the deep map in the 70 and 160
m filter we combined
the two detector time series and projected these together.
The SPIRE data were also reduced using HIPE and maps
were obtained via the default naiveMapper task. The SPIRE observation
consists of several repetitions of a map observation of the same
area. As a result it was possible to project the data with a pixel size of 4, 6, and 9
while still maintaining complete sampling across the source.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.