First release of Apertif imaging survey data

(Abridged) Apertif is a phased-array feed system for WSRT, providing forty instantaneous beams over 300 MHz of bandwidth. A dedicated survey program started on 1 July 2019, with the last observations taken on 28 February 2022. We describe the release of data products from the first year of survey operations, through 30 June 2020. We focus on defining quality control metrics for the processed data products. The Apertif imaging pipeline, Apercal, automatically produces non-primary beam corrected continuum images, polarization images and cubes, and uncleaned spectral line and dirty beam cubes for each beam of an Apertif imaging observation. For this release, processed data products are considered on a beam-by-beam basis within an observation. We validate the continuum images by using metrics that identify deviations from Gaussian noise in the residual images. If the continuum image passes validation, we release all processed data products for a given beam. We apply further validation to the polarization and line data products. We release all raw observational data from the first year of survey observations, for a total of 221 observations of 160 independent target fields, covering approximately one thousand square degrees of sky. Images and cubes are released on a per beam basis, and 3374 beams are released. The median noise in the continuum images is 41.4 uJy/bm, with a slightly lower median noise of 36.9 uJy/bm in the Stokes V polarization image. The median angular resolution is 11.6"/sin(Dec). The median noise for all line cubes, with a spectral resolution of 36.6 kHz, is 1.6 mJy/bm, corresponding to a 3-sigma HI column density sensitivity of 1.8 x 10^20 atoms cm^-2 over 20 km/s (for a median angular resolution of 24"x 15"). We also provide primary beam images for each individual Apertif compound beam. The data are made accessible using a Virtual Observatory interface.


Introduction
Large, untargeted surveys have always been a driver of new discoveries in astronomy, across all wavelength bands. Surveying wide areas of sky or large volumes of the Universe provides large samples of astronomical sources, including many rare and interesting objects.
In the radio regime, surveys have generally fallen into two categories in the past, based on technological limitations. In the first category, radio continuum and polarization surveys utilize interferometers for higher angular resolution with backends that provide a large bandwidth but limited spectral resolution. The prototypical example of a such survey that continues to provide a rich legacy dataset is the NRAO VLA Sky Survey (NVSS) (Condon et al. 1998). The subsequent VLA Faint Images of the Radio Sky at Twenty-Centimeters (FIRST) survey demonstrates the power of even higher angular resolution (White et al. 1997). More recently the VLA Synoptic Sky Survey at slightly higher frequencies is pushing to higher angular resolution and better sensitivity (Lacy et al. 2020), while the Rapid ASKAP Contin-uum Survey (McConnell et al. 2020;Hale et al. 2021), at slightly lower frequencies, is increasing the resolution and sensitivity of radio surveys available in the Southern Hemisphere. Complementary surveys at low frequencies provide an important view of the radio continuum sky. An early example is the Westerbork Northern Sky Survey (WENSS; Rengelink et al. 1997). Recently, the low-frequency window is expanding with a wealth of new low-frequency radio continuum and polarization surveys, including the LOFAR Two-metre Sky Survey (LoTSS) (Shimwell et al. 2019), the LOFAR LBA Sky Survey (LoLSS de Gasperin et al. 2021), and the GaLactic and Extragalactic All-Sky MWA Survey (GLEAM Wayth et al. 2015).
In the second category, spectral line surveys, notably of neutral hydrogen (H i) at 21cm, require high spectral resolution. The largest untargeted H i surveys to date rely on single-dish telescopes, lacking spatial resolution, both to provide sensitivity and backends with the required high spectral resolution over relatively large bandwidths, both of which are necessary to enable effective surveys. The state of the art for H i surveys is the Article number, page 1 of 19 arXiv:2208.05348v2 [astro-ph.IM] 22 Nov 2022 A&A proofs: manuscript no. apertifdr1 Arecibo Legacy Fast ALFA H i survey (ALFALFA Haynes et al. 2018), with the H i Parkes All-Sky Survey (HIPASS Meyer et al. 2004) providing a complementary view in the Southern Hemisphere.
While historically radio continuum and polarization surveys have been distinct from spectral line surveys, the new generation of backends for interferometric radio telescopes can handle large bandwidths at high spectral resolution. The large bandwidth provides for sensitive radio continuum and polarization data while the high spectral resolution enables spectral line observations. Another key technological advance is the development of phased-array feeds (PAFs), which provide a large number of beams on the sky, and hence large fields of view. Thus, a new generation of interferometers operating at GHz frequencies provide the spatial resolution, sensitivity, and field-of-view necessary to efficiently carry out cutting-edge surveys both for H i and for radio continuum and polarization (Maddox et al. 2016). Examples of this include: the MeerKAT International Gi-gaHertz Tiered Extragalactic Exploration (MIGHTEE) survey with MeerKAT, originally designed as continuum survey but which is simultaneously a sensitive H i survey Maddox et al. 2021); and the synergy between the surveys on ASKAP, where the H i-focused Widefield ASKAP Lband Legacy All-sky Blind surveY (WALLABY) (Koribalski et al. 2020) provides a higher frequency measurement to complement the lower frequency the radio continuum and polarization surveys, Evolutionary Map of the Universe (EMU, Norris et al. 2021) and Polarisation Sky Survey of the Universe's Magnetism (POSSUM, Gaensler et al. 2010). This synergy between different observing frequencies can also extend to looking at spectral lines in both absorption and emission, such as between WAL-LABY and the First Large Absorption Survey in H i (FLASH, Allison et al. 2022).
Another example is the PAF upgrade to the Westerbork Synthesis Radio Telescope (WSRT), the APERture Tile In Focus array (Apertif), fully described in van Cappellen et al. (2022). Apertif is a complex system; here, we briefly summarize the most relevant aspects from the perspective of an imaging science user. The Apertif system consists of PAFs with 121 Vivaldi elements on 12 of the 14 WSRT dishes, including dishes RTC and RTD which provide the longest baselines (and highest angular resolution). The signal from these elements can be combined to form forty simultaneous beams on the sky, each with a typical half-power size of 35 , significantly increasing the instantaneous field of view of WSRT to ∼8 deg 2 at 50% of the peak sensitivity. The layout of the beams is shown in figure 25 of van Cappellen et al. (2022). The Apertif system provides 300 MHz of contiguous bandwidth within the frequency range 1130-1750 MHz. At the dish level, the 300 MHz bandwidth is divided into 384 subbands of 781.25 kHz each. The backend can be configured to provide an imaging correlator which further divides each subband into 64 channels of 12.2 kHz.
Following this upgrade, WSRT-Apertif was dedicated to large survey programs: a time-domain survey (Maan & van Leeuwen 2017, van Leeuwen et al. in prep) and a two-tiered imaging survey (Hess et al. in prep). The legacy of a survey is greatly enhanced by the release of high quality science-ready data products to the community. Toward that end, we present the first release of Apertif imaging survey data, covering the first year of survey operations, 1 July 2019 -30 June 2020. In this first data release, the emphasis is on releasing the best quality data products from a first processing to the community as rapidly as possible to demonstrate the potential of the Apertif imaging surveys.
The paper is organized as follows: Section 2 introduces the Apertif imaging surveys, and the observations covered by this release. Section 3 describes the released data products, including the choice for which of these to release. The quality of the continuum, polarization, and spectral line data is discussed in Sections 4, 5, and 6. In Section 7 we detail known caveats and limitations of the Apertif system and data, and we strongly encourage the reader to carefully read this section. We highlight the science potential of the Apertif imaging surveys in Section 8. In Section 9, we briefly describe future prospects and improvements for further data releases. Finally, Section 10 offers a brief summary of the data release. Appendix A provides information on available data from pre-survey operations, including a science verification campaign (SVC) undertaken to verify the scientific performance of the Apertif system and early science observations between the SVC and start of survey operations. A companion paper, Kutkin et al. (2022, hereafter K22), presents a continuum source catalog for the first data release.

The Apertif imaging surveys
The wide field-of-view of Apertif, the spectral resolution of 12.2 kHz across a large bandwidth of 300 MHz, and an angular resolution of up to 11 /sin δ enable large-scale, simultaneous surveys of the radio continuum, polarization, and spectral line sky. Toward this end, a significant fraction of the Apertif observing time has been dedicated to a legacy survey program consisting of two tiers of imaging surveys: a wide and medium-deep tier, referred to respectively as the Apertif Wide-area Extragalactic Survey (AWES) and the Apertif Medium-deep Extragalactic Survey (AMES). The wide tier aims to cover a relatively large region of the sky, while the medium-deep tier focuses on a smaller region of sky, visiting individual fields up to ten times in order to build up sensitivity, with the ultimate goal of achieving a column density sensitivity of ∼5×10 19 atoms cm −2 in order to detect low surface brightness H i. The repeated observations of the mediumdeep tier also enable variability studies.
The design of the imaging surveys was driven by a broad set of extragalactic science goals, including: -Investigate gas and total dynamical mass distribution (derived from gas kinematics) over a wide range of galaxy properties, using resolved H i observations -Investigate environment and the role of interactions, gas accretion and removal of gas -Investigate the properties of the smallest gas-rich galaxies in the local Universe -Study the role of cold gas in active galactic nuclei (AGN) and their feedback activity -Study the history of star formation and AGN activity of the faint radio continuum population -Study the magnetic fields in galaxies and of the large-scale structures in which they are embedded With these goals in mind, the footprint of the Apertif imaging surveys was selected to maximize multiwavelength coverage, especially optical redshifts to provide distances for radio continuum sources. Other considerations included maximizing overlap with LoTSS (Shimwell et al. 2019), providing a large coverage in Right Ascension, and maximizing scheduling feasibility and efficiency. Figure 1 provides an overview of the imaging survey footprint. The full sky coverage of the original four-year survey plan is shown for reference; filled circles indicate the effective final coverage of the Apertif imaging surveys (colored circles are contained within this data release). Fields that have mediumdeep coverage in the full survey are also highlighted in Figure  1. There are two main regions of interest for the medium-deep tier: the Perseus-Pisces supercluster, a rich environment, and the Herschel-ATLAS North Galactic Pole field (Smith et al. 2017), which has large amounts of ancillary data available. In addition, there are two individual fields hosting intra-hour variables with medium-deep coverage. Hess et al. (in prep) provide a detailed discussion on the science cases, chosen footprint, and tiling strategy.
During the first year of survey observations, medium-deep observations were focused on the Perseus-Pisces field. An additional medium-deep field containing the first intra-hour variable found in the Apertif imaging surveys ) was also included. The wide observations focused on coverage between 10 − 16h R.A. with Decl. > 45 • to maximize overlap with publicly available LoTSS data. Wide observations in the same R.A. range but at lower Decl. were also carried out to start building coverage around the second medium-deep footprint. The wide tier in the first year also included significant coverage at 22 − 0h R.A. for maximizing schedule efficiency.
The two tiers share a common observing strategy and observational setup. Survey fields are observed for 11.5 hours as a compromise between flexibility in scheduling and the depth/maximum uv coverage of a full 12 hour synthesis imaging track 1 . Calibrator observations alternate between a flux/bandpass (3C147 or 3C196, occasionally 3C295) and polarization calibrator (3C286 or 3C138); calibrators are observed in every Apertif beam, necessitating forty separate calibrator observations. The calibrator scans vary from 3 − 5 minutes in length per compound beam, depending upon the available gap until the next survey field. Up to two survey fields can be observed in a row; thus a survey field may be separated in time from its calibrator by up to ∼15 hours. Occasionally an observing session will start or end with a survey field in order to increase observing efficiency; then the separation in time may be longer. In order to validate the long time separation between calibrators and target fields, we checked calibration gain solutions over the course of a full observing run (10 days). The amplitude solutions varied by less than 10% over the full period, demonstrating that a longer period between calibration scans should have minimal impact on the absolute calibration. Hess et al. (in prep) describes the automation and details of the survey scheduling. As the main spectral line of interest is H i, the observations use the lowest frequency settings, covering the band 1130-1430 MHz, corresponding to a maximum redshift of 0.26 for H i. This band offers a large cosmic volume for observing H i; however, the lower half of the band is significantly affected by RFI and thus no derived data products are offered for it (see Section 3.3.2) 2 . The 24576 channels provide a frequency resolution of 12.2 kHz, or 2.6-3.2 km s −1 across the band for H i, but spectral line data products are typically produced with a spectral binning to 36.6 kHz. The angular resolution of the imaging survey observations is 11-15 ×11-15 /sin(δ), as the exact resolution depends on the weighting used. Continuum and polarization data are imaged with uniform weighting to provide the highest angular resolution, close to 11 , while the line cubes are produced with a resolution of ∼15 and increased surface brightness sensitivity. We emphasize that the two outlying dishes, RTC and RTD, are required for full angular resolution. The achieved noise in a single observation is ∼40 µJy beam −1 in the continuum images and ∼1.6 mJy beam −1 over 36.6 kHz in the line cubes.

Data release
This section describes the released Apertif data products. All data from the first year of survey observations are considered for release. This includes all raw observational data, and a selection of processed data. In order to balance data quality against the timeliness of a prompt release, processed data products are only released if the continuum image passes validation as described in Section 4.1. This decision is further motivated below in Section 3.3.2. An overview, including the original size, of all Apertif data products is provided in Table 1.

Apercal processing
Survey observations are automatically processed by Apercal 3 , the Apertif imaging and calibration pipeline, via an automated trigger, autocal. Full details of this pipeline are provided in Adebahr et al. (2022b). A brief description of the processing is provided below.
The Apercal pipeline is a modular pipeline written in python using common astronomical software. Apercal processes each beam of an Apertif observation individually, following standard reduction procedures. First, the data is split to only keep frequency ranges mostly free of radio frequency interference (RFI). Experience showed that removing frequency ranges dominated by RFI significantly improved the calibration and quality of the data at all frequencies. For the data under consideration for the data release, this meant splitting to keep the upper 150 MHz of the band, 1280-1430 MHz, and additionally flagging the first 12.5 MHz, in order to avoid persistent RFI, resulting in an effective frequency range of 1292.5 − 1430 MHz. The data is then flagged, both for a-priori known flags 4 and for RFI using AOFlagger, including special adaptations for Apertif (Offringa et al. 2012). A standard cross-calibration (including of crosshand data products) is then done using CASA tasks (McMullin et al. 2007). A direction-independent self-calibration is undertaken in miriad (Sault et al. 1995) on data frequency-averaged to the subband resolution. As a first step, a phase-only parametric self-calibration based on FIRST and NVSS is undertaken; then the self-calibration iterations on the data are undertaken. Phase solutions are always found and applied to the data; amplitude solutions are always derived but only applied to the data if the subsequent dirty image with amplitude solutions improves on the dirty image with only phase solutions. Importantly, Apercal does not account for any direction dependent errors. Finally, the self-calibration solutions are applied and the imaging of the continuum, polarization and line data is done in miriad. For the continuum images, sources above 5-σ are identified and cleaned to a threshold of 1-σ, where σ is the "theoretical" noise, calculated as the noise in a Stokes V image. The same masks are applied to the polarization data and sources there are also cleaned to a 1-σ threshold. The model of the continuum emission is subtracted from the full spectral resolution data and line cubes are created 5 . We note that the line cubes are only imaged, meaning that emission is neither identified nor cleaned in them; for this reason the dirty beam cubes are also kept as a data product to enable later cleaning. Finally, the processed data products are automatically ingested back into the Apertif Long-Term Archive (ALTA). 5 As discussed in Section 6.2, over 75% of released cubes show no (or extremely minor) artifacts; of the remainder, 15% show significant artifacts. These artifacts include incomplete continuum subtraction, in addition to other data quality issues. Thus the single step of visibilitybased continuum subtraction is sufficient in the majority of cases.

Virtual Observatory interface
The released data products are exposed through standard Virtual Observatory (VO) protocols to facilitate their access and exploration to both general and specialized users. In particular, the protocols offered are the Tabular Access Protocol (TAP) 6 and the Simple Image Access protocol (SIA) 7 .
The VO registry can be accessed also through a web browser 8 , which presents all collections on the ASTRON VO registry, including APERTIF_DR1. Selecting one collection allows the user to interactively query through a form all data prod- ucts of that given collection, which is sorted by data product type for Apertif DR1.
The above query capabilities are also possible through standard VO tools, such as using TOPCAT (Taylor 2005) via TAP. The table names to query the different DR1 collections are summarized in Table 2. Moreover, it is possible to query all the available dataproducts from Apertif DR1 at once by using the table ivoa.obscore and limiting the results to the apertif_dr1 by appending to the ADQL statement "where obs_collection="apertif_dr1"".

Extent of data release
All raw observational data from the first year of survey operations are released. A subset of processed data products, meeting specified quality assessment criteria, are also released. These two subsets of released data are discussed further below.

Raw observational data
From the first year of survey observations, there are 221 observations of 160 unique survey fields; with an effective field-ofview of 6.44 square degrees 9 , these observations cover over 1000 square degrees of sky. There are eight medium-deep fields with multiple observations within this data release (64 observations in total, not evenly distributed among the fields); these fields are highlighted in Figure 1.
The raw observational data is recorded in measurement-set (MS) format. The sizes of the original raw data are given in Table 1. A set of calibration scans consists of 40 separate observations taken in succession. For each calibrator scan only one beam contains the calibrator; all other beams not containing the calibrator are discarded to save on data volume. We note that the calibrator observations are taken at a higher time resolution than the survey fields (10 vs. 30 seconds) to allow better RFI excision due to their shorter integration time. 9 Based on a spacing between fields of 2.3 • in Dec. and 2.8 • in R.A.
As can be seen in Table 2, the raw visibility data are provided as separate collections for the survey fields, flux calibrator and polarization calibrator observations. The tables can be joined to identify the associated calibrators for a target observation.
The raw observational data are not stored on disk but rather on tape at a national facility, SURFsara 10 . In addition, the raw data are currently being compressed to further save on data volume (Offringa 2016). Thus, the VO tools enable queries of these data products, but not direct access or download. The data release documentation 11 contains up-to-date information on how to access the data.

Processed data
Processed data products that can be used directly, or with minimum additional processing, are of the most interest to the broader scientific community. The choice to release processed data products is done on a beam-by-beam basis, based on the quality of the continuum image. The continuum data is used for the self-calibration, so an artifact-free continuum image is a good indication that the inherent data quality of the observation and the calibration are of a sufficient standard. This is also a practical choice as the structure of the ingested processed data in ALTA means that either all or no processed data for a given beam should be released. Thus, in order for processed data products for a given beam to be released, the continuum image must pass validation. This approach offers a compromise between releasing as much data as possible on a short timescale to the community while only releasing data that is of high enough quality to be of sufficient interest to the community.
Toward this end, we define quality metrics for continuum images, polarization images and cubes, and spectral line cubes in Sections 4.1, 5.1 and 6.1. The metrics are focused on ensuring a minimum sensitivity and resolution, in addition to identifying significant artifacts in images. The continuum validation criteria in Section 4.1 must be satisfied for processed data products to A&A proofs: manuscript no. apertifdr1  Notes. a The URL are expressed relative to https://vo.astron.nl/apertif_dr1/q be released. The criteria for polarization and spectral line data products do not need to be satisfied, but the data are released with clear flags that indicate if they fail; the frequency of this is discussed in Sections 5.2 and 6.2. We additionally provide the metrics for all data products in the relevant VO tables. A total of 7640 beams were considered for release; 3374 of these beams are included in this first data release. We note that the unreleased beams will be contained in future data releases with improved processing (see Section 9). The sky coverage of the released continuum images is available as a HiPS image 12 .
There are two additional important notes that are relevant for these data products. First, processed data products are only available for the upper 150 MHz of the band; thus the processed data products are produced over the range 1280-1430 MHz. Moreover, the first 12.5 MHz of data are flagged due to persistent RFI, therefore the resulting central frequency is 1361.25 MHz. The nominal bandwidth is then 137.5 MHz, but effectively it could be smaller due to additional RFI flagging. Secondly, these data products are not primary-beam corrected; primary-beam images are provided separately (see Section 3.4); these may be used for mosaicking or for correction of individual images.
The available processed data products are briefly discussed below. The quality of the continuum images, polarization images and cubes, and spectral line cubes are discussed in more detail in Sections 4, 5, and 6.
-Calibrated visibility data Calibrated visibility data, with cross-calibration and self-calibration solutions applied, are currently stored as an intermediate data product at full time and spectral resolution; this may change with future updates to save storage space. These are stored on tape with the raw data, and the data release documentation 13 should be consulted for access.

-Continuum images A multi-frequency synthesis (mfs)
Stokes I image is created over the full frequency range (1292.5 -1430 MHz) and saved as a fits file for each beam. The size of the continuum images is 3.4 • ×3.4 • (3073×3073 pixels, with 4 /pixel). This extends well outside the primary beam response to account for cases where a strong source is in a sidelobe and needs to be included in the self-calibration model and cleaning. Sources at the 5-σ level should be identified and cleaned to the 1−σ level 14 .  Table 3 summarizes the covered frequency ranges and provides the corresponding redshift range for H i. The lowest redshift cube is produced at full spectral resolution while other cubes are produced with a 3-channel averaging. These cubes have a spatial extent of 1.1 • ×1.1 • (661×661 pixels, with 6 /pixel). As the Apercal pipeline does not provide source finding or cleaning of the line cubes, corresponding dirty beam cubes are also provided to allow offline cleaning of source emission. Originally these dirty beam cubes are produced with twice the spatial coverage of the line cubes to enable full cleaning 15 . These dirty beams cubes are provided as linked data products in the spectral cube VO table.

Related data products: Primary beam images
Important additional data products are the primary beam images for each of the forty Apertif beams. Currently, Apercal does not provide primary beam corrections to any data products. Instead, primary beam corrections are applied separately, which requires primary beam images. Due to the nature of the phased array feed, each Apertif compound beam has a unique primary beam response and thus primary beam images are required for each of the forty beams. As Apertif does not offer a holography mode, we undertake two different approaches for measuring the primary beam response: drift scans across extremely bright sources (Dénes et al. 2022) and reconstruction of the primary beam responses via comparison to the NVSS catalog with a Gaussian 2022b), if there are persistent artifacts in a residual image, the maximum number of clean cycles may be reached before sources down to the 5-σ level are included in the source mask. 15 We take advantage of the symmetric nature of the WSRT restoring beam to trim the beam cubes in half around the North-South axis to halve the size of the beam cubes. A script and instructions for restoring the full beams cubes are available at https://github.com/ apertif/trim_apertif_beam_cube  Notes. a Optical velocity definition, with v = cz b For center of cube process regression technique (K22). For this first data release, the primary beams from the Gaussian process regression approach are released. K22 describes their derivation in detail, and the data release documentation contains information on how to best apply these primary beam images to the Apertif data, including how to scale the primary beam images with frequency.

Continuum image quality
This section describes the validation of the continuum images and provides an overview of the data quality. In addition, we report on the accuracy of the flux scale and astrometric precision.

Validation of continuum images
The continuum images of every beam are individually validated. The starting point of the validation is the residual images obtained after cleaning the continuum images, and the validation aims at checking to what extent these images contain only Gaussian noise. The premise is that any significant deviation from Gaussian noise in the residual image indicates issues with the calibration and reduction of the data. The following parameters were derived for each residual image: σ in : Noise in the inner half degree of the image 16 .
σ out : Noise at the edge of the residual image, more than a degree from the centre. where pixels with values below this level cover 10 beam areas within the inner 0.5 • of the residual image.
Both σ in and σ out are determined in a robust way using the median of the absolute values. σ in is the relevant noise for scientific analysis, including the contributions of imaging artifacts, while σ out is a reasonable measure of the expected noise. b min corresponds to the angular resolution in an image; it is chosen over the corresponding size of the major axis of the restoring beam (b max ) as b min depends only on the baselines available for an observation, while b max depends also on the declination of an observation. R is a measure of the strength of artifacts left in the center of the residual image. Ex−2 is a measure of how much a residual image deviates from perfect Gaussian noise; in the case of perfect noise Ex−2=0. This metric is derived by comparing 16 Specifically, within the central 500×500 pixels to a Gaussian corresponding to σ out . Figures 2 and 3 demonstrate the actual noise histograms and comparison Gaussian distributions. Neg10 is another measure of how much a residual image deviates from Gaussianity. We emphasize that it is defined as an absolute level, so the expected value for a pure Gaussian distribution is 3.2, and more positive values indicate significant (negative) calibration residuals. We note that we did not use the equivalent of the parameter Ex-2 or Neg10 based on positive deviations from Gaussianity (Ex+2, Pos10). This is because insufficient cleaning results in source emission remaining in the residual image which would then dominate the validation (see the upper right panel of Figures 2 and 3).
We then set the following criteria to pass validation; in order for a continuum image to pass validation and be released its residual image must satisfy all four of the following requirements: 1. σ in < 60 µJy/beam 17 . 2. b min <15 . 3. R < 1.225. 4. R < 1.15 or Neg10 < 4.5 or Ex − 2 < 400.
The first two criteria were chosen to ensure that minimum survey specifications were met. Criterion 1 ensures that the noise of the images is low enough to be considered survey quality and valid. Criterion 2 ensures a minimum angular resolution that is required to achieve the science goals and meet survey specification. The restoring beam can be larger when both dishes with the longest baselines (RTC and RTD) are missing from an observation, and observations that are missing both RTC and RTD are considered failed 18 .
The second two criteria were set to ensure a minimum of imaging artifacts in continuum images. These criteria were determined by undertaking visual examination of a large set of images to determine numerical criteria that would catch significant image artifacts. The main types of image artifacts are due to errors in the self-calibration as well as strong directiondependent errors for which the current calibration pipeline does not attempt to correct. Criterion 3 ensures that images do not have too much extra noise in the inner parts of the residual image, corresponding to left-over artifacts. Criterion 4 focuses on edges cases, where R is at intermediate values and additional constraints on large-scale structures and deviation from Gaussianity in the residual image are needed. We note that the first criterion can also eliminate images with significant artifacts if these increase the noise in the inner part of the image above the threshold. The criteria were set so that the large majority of images which were visually assessed as good would pass while 17 Since σ in is always greater than σ out this is also a criterion on σ out 18 Such observations are repeated for individual wide fields while for medium-deep fields the eventual combination with data that contain RTC and RTD provides the final required angular resolution A&A proofs: manuscript no. apertifdr1 These include the metrics in Section 4.1, in addition to other statistics not (currently) used in the validation. The first five are: σ in , σ out , the peak in the continuum image, the restoring beam, and R. The next two are Ex−2 and Ex+2. The dynamic range (D) is calculated as the peak of the continuum image divided by Neg10×σ out . The final two statistics are Neg10 and Pos10. minimizing the number of images visually assessed as bad which would pass. Figures 2 and 3 provide an example of the diagnostic plots and validation metrics for two beams, one of which fails validation and one which passes. The image that fails validation has σ in = 55.7 µJy beam −1 , b min = 11.4 , R = 1.24, Ex−2 = 466, and Neg10 = 5.5; thus is fails both criteria 3 and 4. The validation metrics are provided for all released continuum images in the VO table.  Table 4.

Overview of released continuum image quality
The behavior of different compound beams is not identical. Specifically, the outer compound beams illuminate the edge of the field of view and thus may be expected to have a reduced sensitivity. This is demonstrated in figure 38 of van Cappellen et al. (2022) which shows the noise of continuum images in the compound beam layout, normalized to the beam with the lowest average noise, which is beam 24. Beams closer to the edge of the field of view have higher continuum noise values.

Accuracy of flux scale
K22 assess the accuracy of the flux scale. They cross match the Apertif continuum catalog to the NVSS catalog and find that the flux scales match, which is by design of the primary beam derivation. They find a significant scatter in the flux ratios, which is a combination of multiple effects, including long-term variability of sources and the different angular resolution and sensitivities of the two catalogs. In order to assess the precision of the Apertif flux measurements, K22 compare flux measurements for medium-deep epochs with at least five observations, and find that flux variations across time indicate that the measurement precision is better than 10%, consistent with expected errors from cross-calibration.

Astrometric accuracy
K22 perform a cross-match of the Apertif radio continuum catalog to the public LoTSS catalog (Shimwell et al. 2022). As part of this cross-match they confirm the astrometry of the Apertif radio continuum images, finding that the typical offset between the Apertif and LoTSS sources is 0 with a standard deviation of 2 in both the R.A. and Decl. directions.
We note that the self-calibration of the Apertif data begins with a parametric self-calibration using FIRST or NVSS data; thus the Apertif data is calibrated to the FIRST and NVSS astrometry by design.

Polarization data quality
Compared to the continuum images, there are a few important points to keep in mind for the polarization data quality. First, due to the physical nature of Stokes Q, U and V, astrophysical emission in the polarization images and cubes can be a positive or negative quantity while it can only be positive in Stokes I. Secondly, for Stokes Q and U cubes even faint artifacts in individual images can stack up, if present over the whole cube at a similar position, when the rotation measure synthesis technique is applied. This is also possible vice versa where strong artifacts in an individual image plane can be averaged out by rotation measure synthesis. Finally, Stokes V represents the circular polariza-Article number, page 9 of 19 A&A proofs: manuscript no. apertifdr1 Notes. a For all released beams, which is based on the continuum image passing validation b For only those beams which pass validation for the specified data product (good quality for line cubes, see Section 6.1); for continuum images this is the same as the released beams, while it is a subset for polarization and line.  tion. Astronomically circularly polarized sources are extremely rare and most often show polarization fractions below 1%, so that Stokes V images should nominally be regarded as empty of astrophysical signal.

Validation of polarization images and cubes
After the overall calibration of the data, the quality of the polarization images and cubes is mostly influenced by the instrumental leakage characteristics of the primary beam. This means that the strongest artifacts often appear for sources far away from the beam centres where the instrumental leakage is higher. With this in mind we defined the following metrics, in addition to those in Section 4.1: The validation of the polarization images and cubes provides information on their data quality; however, the release of polarization data products is based on the validation of the corre- sponding continuum image (Section 4.1), as detailed in Section 3.3.2. The Stokes V image and Stokes Q and U cubes have separate validation criteria, due to their different nature.

Validation of Stokes V images
In order for a Stokes V image to pass validation (as indicated by the data quality flag in the VO table), the following four criteria must all be satisfied: 1. σ in and σ out < 60 µJy/beam. 2. b min < 15 . 3. FT max < 25. 4. p in < 4 mJy.
The first two criteria were chosen to ensure minimum survey specifications were met and mirror the first two criteria for continuum images. Thus, generally a released Stokes V image will naturally pass those criteria as the (residual) continuum image has also passed them. Criterion 1 ensures a minimum sensitivity in the polarization images, and criterion 2 ensures the angular resolution.
The second two criteria were set to ensure a minimum of imaging artifacts in the Stokes V images. These criteria were determined by undertaking visual examination of a large set of Stokes V images. In addition to artifacts from errors in selfcalibration and direction-dependent errors (which are caught by the continuum image validation, required for data release), leakage of strong sources near the primary beam edges also play a large role. In particular, criterion 4 eliminates images with strong instrumental leakage. Criterion 3 is designed to eliminate images with stripes due to unflagged RFI or bad amplitude selfcalibration solutions. The criteria were set so that the large majority of images which were visually assessed as good would pass validation while only a small fraction of images that were visually assessed as bad would pass.

Validation of the Stokes Q and U cubes
In order for Stokes Q and U cubes to pass validation (as indicated by the data quality flag in the VO table), at least two-thirds of the image planes within the cubes must satisfy both of the following criteria: 1. b min < 17.5 . 2. σ in and σ out < 300 µJy/beam.
The resolution requirement in criterion 1 is relaxed slightly compared to the continuum or Stokes V images due to the fact that the restoring beam becomes larger at lower frequencies. The noise requirement in criterion 2 is consistent with the requirement on the Stokes V images above, assuming 24 individual images covering the same total bandwidth, giving a factor of ∼5 increase in noise.

Overview of released polarization data quality
The validation of the polarization images and cubes is not required for release but is undertaken to provide information on their data quality. Generally, the validation of the polarization images and cubes follows that of the continuum very closely. This is due to the use of similar metrics and the fact that the continuum validation already excludes the majority of polarization images and cubes that would have large artifacts. Of the released 3374 beams, which are required to pass continuum validation (see Sections 3.3.2 and 4.1), only 155 fail the Stokes V validation and 58 fail the Stokes Q and U validation and are flagged as such in the VO table. Generally, the reason a polarized data product might fail validation while the continuum image does not is due to a strong source at the edges of the image, where the leakage is largest. The validation status and metrics of the polarized data products are provided in the relevant VO table (see Table 2). In addition, there are 21 released beams that have no polarized data products because a polarized calibrator was unavailable for the observation 200309042 (S1042+5324).
The median inner and outer noise values across all Stokes V images are presented in Table 4, where the uncertainties indicate the 16 th and 84 th percentiles. There is essentially no difference in median noise values when considering all released Stokes V images, or only those that pass validation, supporting the release strategy outlined in Section 3.3.2 based on continuum image validation only.
In addition to the metrics presented here, Adebahr et al. (2022a) use polarization data from the SVC to place a limit of 1% on the polarization leakage down to the 30% response level of the primary beams.

Line cube quality
The quality of line cubes is dominated by different effects than the continuum or polarization data. While imperfect calibration or direction-dependent effects can impact the spectral line cubes, generally artifacts in the line cubes are dominated by two categories: imperfect continuum subtraction or bad frequency ranges (e.g., imperfect sub-bands at the system level).
It is useful to recall that four different line cubes are produced by Apercal; see Table 3 for a summary of their properties. The highest frequency cube (lowest redshift), cube3, is produced at the highest spectral resolution. Thus, it will have different noise properties compared to the other cubes. The RFI environment generally worsens at lower frequencies, so cube0 will be more strongly affected by RFI than cube1, for example.

Validation of line cube quality
The validation of the line cubes focused on the three lower spectral resolution cubes, as a common set of metrics could be defined for them. In practice, the quality of the high spectral resolution, high frequency cube3 follows that of the neighboring (in frequency) cube2. The following metrics were defined for each of cube0, cube1 and cube2: σ: The rms noise f ex : Fraction of the total number of pixels with an absolute value > 6.75σ p 0.8 : Ratio between the width of the noise histogram at a level of 0.8% of the maximum and σ The last two metrics were empirically determined to be successful at identifying deviations from Gaussianity in the noise histograms, which correspond to artifacts in the cubes (see Figure 6).
Since the line cube quality can differ significantly from the continuum quality, upon which release of processed data is based (see Sections 3.3.2 and 4.1), this first data release can contain cubes of poorer quality. Given this, rather than a strict pass/fail validation, a slightly more nuanced system of "good", "okay" and "bad" was adopted for line cube quality: -Good: No (or very minor) artifacts -Okay: Minor artifacts present but would not significantly impact analysis -Bad: Major artifacts that have to be accounted for in analysis "Good" cubes satisfy all three of the following criteria: 1. σ < 3 mJy, 2. log( f ex ) < −5.30, 3. p 0.8 < 0.25 f ex + 5.875.
"Bad" cubes are those that do not meet the "good" or "okay" criteria. Using these conditions, we assigned all of cube0, cube1 and cube2 a ranking of "good", "okay" or "bad". Cube3 was assigned the quality of cube2 in all cases.
In order to determine the above criteria, a large number of cubes (∼ 550 cubes from 14 observations) were visually inspected. Figure 6 demonstrates the inspection plots, including the noise histogram, an image of a representative channel, and a position-velocity slice. Generally, the first criterion ensures a minimum sensitivity while the second two criteria identify deviations from Gaussianity in the noise distributions of the cubes.
A&A proofs: manuscript no. apertifdr1 Fig. 6. Examples of the three quality classes used for the HI quality assessment. The top row shows an example of a "good" observation (ObsID 190914041, beam 5, cube 2), the middle one an "okay" observation (ObsID 190914041, beam 8, cube 2) and the bottom one a "bad" observation (ObsID 190914041, beam 3, cube 2). The columns show, from left to right, the noise histogram, an extract of the central velocity channel, and a position-velocity diagram through the center of the cube (the horizontal lines in the position-velocity diagrams are the subband edges; see Section7.2 ). In the left column, the short horizontal line indicates the rms, σ, and the two dotted vertical lines indicate ±6.75σ. The "good" observation in the top row shows hardly any artifacts and a Gaussian noise histogram. The "okay" observation in the middle row shows a minor continuum subtraction artifact, which causes somewhat extended wings to the noise histogram. The "bad" observation in the bottom row shows significant continuum subtraction artifacts, resulting in a very non-Gaussian noise histogram.

Overview of released line cube quality
The validation of the line cubes can diverge quite significantly from that of the continuum images. This is because the two main causes of line cubes failing validation, incomplete continuum subtraction and bad frequency ranges (whether from heavy flagging due to RFI or subbands with poorly measured beamweights), can be independent of the quality of the continuum images and self-calibration 19 . As incomplete continuum subtraction can be addressed with a secondary subtraction in the 19 While poor continuum calibration is likely to lead to worse continuum subtraction, good continuum calibration does not ensure good con-image plane using the line cubes and bad frequency ranges are an unfortunate feature of radio observations, we do not require the line cubes to separately pass validation. Instead, the quality of the data cube ("G"ood, "O"kay, or "B"ad) is clearly indicated in the associated VO table (see Table 2). In total, 10,317 of the released cubes are good, while 1116 are okay, and 2063 are bad. Thus, over 75% of cubes fully pass validation while only 15% are rated as bad. The cubes with a quality assignment of bad may tinuum subtraction. In particular, poor bandpass calibration can lead to poor continuum subtraction. benefit from further off-line processing (i.e., continuum subtraction) before analysis. Table 4 provides a look at the median rms values for the different spectral cubes across all the released cubes; the uncertainties represent the 16 th and 84 th percentiles. There is a trend that the noise values are slightly higher for the lower frequency cubes, consistent with a worsening RFI environment at lower frequencies, in addition to the fact that the system performance is generally worse at lower frequencies. Limiting the assessment to only "good" cubes decreases the median noise values slightly, as would be expected, but the change is minor, thus supporting the strategy of releasing all cubes where the continuum image passes validation.
The median noise of 1.6 mJy beam −1 corresponds to a 3σ column density sensitivity of 1.8×10 20 atoms cm −2 over 20 km s −1 ; this is for a median declination of 38 • for released beams, which corresponds to an angular resolution of 24 ×15 for the spectral line cubes.

Known caveats
As described above, we have undertaken validation of produced data products in order to identify high quality data for release. However, there are a few global issues that affect data quality which we briefly describe below.

Ghosts
The Apertif data suffer from "ghosts" at the center of images: bad signal with random phase (that averages to zero). These ghosts are most prominent in channels 16 and 48 of each subband. While early versions of the Apercal pipeline flagged these channels, this was disabled for all data products presented here. Before the SVC period, the finite impulse response (FIR) filter in the channel filterbank was enabled. This severely reduced the presence of ghosts in channels 16 and 48, but did not completely remove them, and the ghost signal was also present at a low level across all channels. Due to the reduced presence of the ghosts in channels 16 and 48, plus the low level contribution from all channels, the decision was made to not flag channels 16 and 48 of the subbands as the extra bandwidth outweighed the presence of the ghosts. The presence of these ghosts builds up when averaging data over frequency, and so they affect all Apertif data products. Thus, any source identified at the exact center of a beam should be treated with extreme caution.

Aliasing
The coarse channelization of the data into subbands uses a filter that does not have a perfectly sharp frequency response. This results in some overlap of response between adjacent subbands. This effect is strongest for channels near a subband edge and also results in a sharp drop in overall response for channels at the subband edges, namely channels 0, 1 and 63 of every subband. No correction is done for the aliasing, and the pipeline uses a brute force approach of flagging the channels with suppressed signal at the subband edges. Thus, we note that 3 out of every 64 channels are flagged. Since cubes0-2 have a three channel averaging (which does not divide evenly into 64), the presence of these flagged channels shifts around, and the manifestation of this flagging alternates between a single channel that is fully flagged, or two adjacent channels that are partially flagged (1 or

Scientific potential
In this section, we briefly highlight the scientific potential of the data contained in this first data release.

Continuum images
The Apertif continuum images have three times the angular resolution of NVSS and are ∼10 times as sensitive. Thus, the Apertif images will both resolve more source structure and also detect fainter emission than NVSS. The upper two panels of Figure  7 illustrate this, showing a mosaicked Apertif image (K22) and a NVSS image over the same field of view. Many more point sources are seen in the Apertif data, and the excellent surface brightness sensitivity of WSRT-Apertif combined with angular resolution, reveal the features of an extended, diffuse source. The sensitivity and resolution of the Apertif continuum images are comparable to those achieved by the Evolutionary Map of the Universe (EMU) survey with ASKAP (e.g., Gürkan et al. 2022); with complementary sky coverage the two surveys augment each other well.  Figure 7 demonstrates the comparison between Apertif and LoTSS data. The combination Apertif-LOFAR is particularly relevant for the study of the spectral properties of radio sources. The sensitivity of the surveys with the two instruments is well suited to trace the typical spectral index observed in extragalactic sources, around α ∼ 0.7 or steeper 21 . Specifically, given the relatively high spatial resolution, it is possible to derive resolved spectral index images (as illustrated in the bottom right panel of Figure 7) for both single objects and large samples. In particular, significant steepening of the spectrum at low frequencies reveals the presence of a particularly old electron population, possibly indicating remnant emission where the nuclear activity has stopped and the electrons have not been replenished. Morganti et al. (2021a,b) presented searches for structures with such extremely steep spectral indices by combining Apertif and LOFAR 150 MHz images in the Lockman Hole region. Finding samples of these elusive objects allows us to put constraints on their life-cycle (see Jurlin et al. 2020;Morganti et al. 2021b). Resolved spectral index studies are also important for diffuse emission in clusters (Orrù et al. in prep., Shulevski et al. in prep), for single objects like the radio galaxy B2 1321+31 (Morganti et al. 2021a) as well as for large major mergers like Mrk 273 (Kukreti et al. 2022). These studies will be further expanded by the combination with the LBA LOFAR survey, as seen in the study of Mrk 273 (Kukreti et al. 2022).
The east-west nature of WSRT means that if a source varies during the course of an observation, it leaves a very clear pattern in the images. Oosterloo et al. (2020)  intra-hour variable during the SVC, which was then monitored as a medium-deep field over the course of full operations. To date, about twenty intra-hour variables have been identified in the Apertif survey data, along with a few radio variable stars (Oosterloo et al. in prep). The image products in this data release will also complement any future independent detections of transients and variables in the area it covers. This has recently been demonstrated for both fast radio bursts and for radio afterglows of gravitational-wave events. Connor et al. (2020) detected a bright fast radio burst in the Apertif time-domain survey , and then used the Apertif continuum image to set limits on any accompanying persistent radio source. Boersma et al. (2021) searched Apertif data for a radio counterpart to binary neutron-star merger gravitational-wave event GW190425. While none was found for that particular event, future searches can use DR1 as the baseline comparison for the radio-afterglow detection.
The companion paper K22 provides a continuum source catalog based on all continuum images contained in this data release. This includes a comparison to NVSS to identify sources that vary on long time scales, plus a cross-match to the LoTSS DR1 to provide spectral indices for unresolved sources.

Polarization images and cubes
The Apertif polarization data, as with the continuum data, provide an improvement of three times the angular resolution and more than ten times the sensitivity when compared to NVSS, the current benchmark for wide surveys of the polarized sky. The sensitivity of the wide tier results in a polarized source density of 21 sources deg −2 (Adebahr et al. 2022a), a factor of 20 improvement compared to NVSS (Taylor et al. 2009). This increased density of polarized sources enables a higher angular resolution rotation measure map to be constructed, thus providing a more detailed look at the magnetic field of our own Milky Way galaxy. The sensitivity of the Apertif surveys also means that the polarized sky is being measured down to the µJy level. An open question is at what level the polarized sky changes from being dominated by AGN to star-forming galaxies; this happens at a lower flux level than for the total power radio continuum sky. Adebahr et al. (2022a) use early Apertif data to show that polarized source counts are still dominated by AGN, but that the AGN are hosted in late-type, rather than early-type, galaxies. Additionally, the high density of polarized sources in the Apertif data provides the possibility to robustly investigate the source counts of polarized sources and behavior of fractional polarization with total intensity (Berger et al. in prep).
The synergy with LOFAR, and LoTSS in particular, is also key for polarization. Having information over a decade in frequency allows not only an investigation of the polarization behaviour with respect to the sources' spectral index, but also an investigation of polarization spectra. Due to Faraday Dispersion being strongly wavelength dependent, depolarisation effects are much more dominant at longer wavelengths. Combining LOFAR and Apertif data thus allows a deep insight into the nature of depolarisation mechanisms and their physical origins (Berger et al. in prep). Such an analysis also provides information on the morphology and turbulence of the magnetic fields, not only in the host objects, where the polarised emission is generated, but also along the line-of-sight to the observer. When using additional redshift information this enables an analysis of the evolution of cosmic magnetic fields (Berger et al. 2021).

Spectral line cubes
The Apertif spectral line cubes offer angular resolution combined with an untargeted detection of line emission, independent of the stellar content of galaxies. Compared to single-dish H i surveys, such as the ALFALFA H i survey, the angular resolution improves by > 10×, allowing the distribution of the H i to be studied, providing information on disturbed morphology and environmental effects, resolved kinematics, and more. In terms of sensitivity and angular resolution, the wide tier is well-matched to WHISP, a targeted survey of 375 galaxies with WSRT (van der Hulst et al. 2001). Figure 8 shows UGC 11951 as seen in both Apertif and WHISP, where both datasets show a warp in the H i disk. The Apertif spectral cubes offer WHISP-quality data but for thousands of galaxies, with no selection criteria based on the stellar content. This enables the detection of galaxies with minimal stellar content, and the derivation of their kinematic properties to study their behavior on galaxy scaling relations (Šiljeg et al. in prep). The Apertif H i detections are being used to select galaxies, based on their H i morphology, for observation with the large IFU of the new WEAVE instrument (Jin et al. in prep), as part of the WEAVE-Apertif survey. Figure 9 demonstrates the power of the angular resolution combined with untargeted selection, showing a galaxy with a hugely extended H i disk whose stellar content appears relatively undisturbed.
The Apertif spectral line cubes are not solely H i cubes but OH megamasers can also be detected over their frequency coverage. Hess et al. (2021) report the serendipitous discovery of an OH megamaser in the Apertif spectral line cubes and discuss expectations for the full Apertif imaging surveys, including how the wide bandwidth permits the simultaneous detection or placement of limits on OH megamaser satellite lines in addition to detecting the redshifted main lines.

Future prospects
This is the first of multiple data releases to come from the Apertif imaging surveys. Improvements offered by future data releases can be broken into three broad categories: 1. More sky coverage As the processing of the Apertif imaging survey data continues, there will be more processed data that can naturally be released in the same manner as this data release. 2. Improved data processing A major component of this release was to determine which data products were of high enough quality to be released. Improvements to the data processing will result in more high quality data to release, improving the overall sky coverage of the public data. 3. Higher order data products This release initially provides per-beam images and cubes. A companion paper K22 provides a continuum source catalog. The ultimate goal is to offer more advanced science products to the community. This includes mosaicked images, but also source cutouts and cubelets and source catalogs for all data products (polarization and spectral line). For example, Adebahr et al. (2022a) provide a polarized source catalog for the SVC observations.
Below, we detail a few of the key processing improvements that are in progress, which we plan to incorporate in future data releases.
-Direction dependent calibration: In some images produced with Apercal, there are direction dependent errors (DDEs) remaining around sources. We believe these errors come from two main sources: (1) pointing errors in the dishes and (2) faulty PAF elements leading to specific compound beams on a given dish having a non-ideal beam response that is different from other dishes (see Dénes et al. (2022) for more details). These errors cannot be removed in the standard self-calibration procedure implemented in Apercal.
We took operational steps to address these errors, updating the pointing model and targeted maintenance of faulty elements. However, there will always be some bad elements, and hence compound beams with bad shapes. Thus, we are also working on including direction-dependent calibration in Apercal, where the antenna gains are solved in the directions of bright sources simultaneously, pushing the amplitude of artifacts below the thermal noise level. -Two-stage continuum subtraction The single-stage continuum subtraction is not sufficient in many (∼20%) cases. Implementing a second continuum subtraction step will increase the quality of the final line cubes by minimizing the presence of continuum artifacts. -Subband-edge flagging The lack of an anti-aliasing filter results in suppressed signal on subband edges; the current approach to handling this is to flag three of every sixty-four channels, significantly affecting the final line cubes. This is likely an overly generous flagging approach and only the first channel of every sixty-four needs to be flagged, while the bandpass correction can account for the subband response in the second and last channels of each subband. -Mosaicking We are also working on a mosaicking procedure so that in future data releases, mosaics (both of single Apertif pointings and also across pointings) can be released in addition to individual beam images. K22 provides an overview of a mosaicking tool used in the creation of the continuum catalog that undertakes the different primary beam correction for various Apertif beams, including a link to source code.

Summary
This paper presents the first release of data products from the Apertif imaging surveys, covering the first year of survey observations. The release consists of two major components: the raw . The color scale is the same between the two images and the contours are at [2,4,8,16,32] × 10 20 atoms cm −2 . We note that the WHISP data have a smaller beam size as uniform weighting was used; this can also account for the difference in peak column density values between the two images. The Apertif data is from a single beam; mosaicking would further increase the signal in that map. observational data, plus processed data products. The raw observational data covers 300 MHz of bandwidth with 12.2 kHz resolution and provide an angular resolution up to 11 /sin δ. This data release consists of 221 observations of 160 unique survey fields, corresponding to an effective sky coverage of ∼1000 deg 2 . Processed data products available include continuum images, Stokes V images, Stokes Q and U cubes, and four spectral line cubes plus associated dirty beams. The processed data products have a bandwidth of 137.5 MHz; the lower 162.5 MHz of the band is discarded due to strong RFI. We release all processed data products for beams where the continuum image passes validation, consisting of 3374 observations of 2683 unique fieldbeam combinations. A total of 44% of all possible beams are released. The main reasons for continuum images to fail validation are artifacts from poor self-calibration or direction-dependent errors, both of which will be addressed in future versions of the pipeline.
The metrics and validation status of all data products are provided in the associated VO tables ( Table 2). The continuum images have a median inner noise of 41.4 µJy beam −1 ; the Stokes V images have a median inner noise of 36.9 µJy beam −1 ; and the spectral line cubes have a median noise of 1.6 mJy beam −1 over 36.6 kHz. The median angular resolution of the continuum images and polarization data is 11.6 /sin(δ).
The processed SVC data was released without validation as a demonstrator of data quality, both the intrinsic quality and pipeline processing. For comparison to the survey data release, Table A.1 provides a high-level overview of the number of beams that would have passed continuum validation for each field, if the validation process as for the survey data release had been used.

Appendix A.2: Early science observations
There was a ∼2.5 month period between the SVC and start of survey operations. This period was focused on development and finalizing tools for operations but did offer the possibility for high quality early science observations. These observations are listed in Table A.3. These data were observed at different frequency settings and only the last observations have the online correction for delay tracking. We are releasing these observations but not associated processed data; none of these observations have been fully processed by the Apercal pipeline. We do note that one field (ObsID 190428055) was calibrated with Apercal and manually imaged and published in Morganti et al. (2021b).