Open Access
Issue
A&A
Volume 658, February 2022
Article Number A83
Number of page(s) 19
Section Astronomical instrumentation
DOI https://doi.org/10.1051/0004-6361/202142275
Published online 04 February 2022

© C. B. Adams et al. 2022

Licence Creative CommonsOpen Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1. Introduction

When energetic γ-rays or charged particles (typically protons, atomic nuclei, or electrons) enter the atmosphere, they generate a cascade of secondary particles (also known as an extensive air shower or EAS) through pair production, bremsstrahlung emission, and, for the case of hadronic showers, fragmentation and decay of unstable mesons (π0, π±, and K±). The resulting ultra-relativistic particles, traveling faster than the local speed of light in the atmosphere, produce coherent polarization of the dielectric medium. This in turn produces beamed Cherenkov radiation in the forward direction, forming a light pool at ground level of approximately 150 m radius.

Imaging atmospheric Cherenkov telescopes (IACTs) sample this Cherenkov light pool where the optical assembly of each telescope collects and focuses the light onto a corresponding camera, comprising an array of photo-sensitive detectors. As the shower progresses from the top of the atmosphere to the ground, Cherenkov light is produced with a duration of up to 100 ns. From the ground, camera pixels register the flash of light to form an image of the shower. The shape and time-gradient of that image depends on the properties of the primary particle, on its arrival direction, and on the distance of the air shower to the telescopes. Images due to pure electromagnetic showers tend to be compact, well-defined, and of an approximately elliptical shape. Hadronic showers, on the other hand, tend to generate more penetrating secondary particles (e.g., muons) at higher transverse momenta, resulting in images with more clumpy shapes. Particle shower images, particularly of γ-ray origin, can be described in terms of a first and second moment analysis whose parameters are used to derive the properties of the primary particle (Hillas 1985).

The observed spectrum of Cherenkov light that is generated in extensive air showers extends from about 200 nm to more than 700 nm. The generated Cherenkov light follows a 1/λ2 distribution; absorption and scattering of the light in the atmosphere affects mostly the ultraviolet (UV) part of the spectrum. As a result, the observed flux at ground level from γ-ray showers is significantly reduced below a wavelength of about 280 nm. Even so, the bulk of the observed emission happens at short wavelengths, decreasing in intensity as the wavelength increases. This, in addition to the presence of strong airglow lines and increasing night sky background (NSB) starting from 550 nm, explains why IACTs are designed to be mostly sensitive to blue and near-UV radiation. The number of Cherenkov photons emitted in the shower is approximately proportional to the energy of the primary particle (Hillas 1985; de Naurois & Mazin 2015). This is particularly true for air showers that are generated by primary electrons and γ-rays.

IACTs operate in harsh environments with variable weather conditions, wide temperature ranges, and occasional snow, rain, or dust storms. Nevertheless, these telescopes lack the protective buildings of other optical instruments, such as astronomical domes. The optical components of IACTs, which are designed to collect and focus the Cherenkov light, suffer from degradation processes due to the aforementioned weather conditions. In addition, the cameras of IACTs are usually made of photomultiplier tube (PMT) pixels operating at high voltages and they degrade as the total accumulated charge increases. During standard operations, PMTs are exposed to NSB light, typically inducing pixel currents of 5 − 10 μA, and up to 15 − 20 μA during exceptionally bright moonlight conditions. These factors combined can induce aging that impacts both the optical and electronic response to incoming light, including Cherenkov light from EAS. Monitoring and correcting for the varying instrument performance therefore becomes a key element in the calibration and analysis of Cherenkov data.

Hillas & Patterson (1990) originally proposed to monitor and calibrate IACTs using images generated from local high energy muons generated by hadronic showers. First implemented for the calibration of the Whipple telescopes (Fleury 1991; Rovero et al. 1996), various IACTs have used muon images to study the changing responses of the telescopes (Vacanti et al. 1994; Aharonian et al. 2004; Meyer 2005; Gaug 2006; Chalme-Calvet et al. 2014). The Cherenkov emission generated by these particles is emitted in a cone with a roughly constant opening angle, appearing as a ring when observed from the ground at small incident angles. The radius of the ring depends on the properties of the incident particle (e.g., energy) and is not affected by instrument throughput. It can be compared with the number of photo-electrons seen by the telescopes to provide a measure of the efficiency of the telescope detecting optical light. In practice, the analysis of muon images is challenging for many reasons. It involves a detailed geometrical reconstruction, particularly when part of the muon ring image falls outside the camera’s field of view (FoV), when the shower trajectory is tilted or displaced with respect to the center of the FoV or if groups of pixels are malfunctioning. In addition, as Cherenkov photons from muons are produced close to the telescopes, atmospheric absorption is less severe and the Cherenkov photon spectrum is shifted to shorter wavelengths compared with photons generated by γ-ray showers (Chalme-Calvet et al. 2014). Therefore, the calibration method using the muon images requires additional corrections or specific Monte Carlo (MC) simulations in order to provide an accurate estimation of the throughput for γ-ray showers (Gaug et al. 2019). Due to this added complexity, the Very Energetic Radiation Imaging Telescope Array System (VERITAS) currently uses muons mostly as a supporting technique to monitor the calibration workflow (see Sect. 2.3.1).

This work discusses a different method to measure the total throughput. It is based on the characterization of mirror reflectivity and the continuous monitoring of camera gains. The relative variations of these parameters are used to correct the simulated γ-ray event pulse charges to provide throughput-corrected response functions. We discuss how this method was successfully implemented in VERITAS, providing a reliable way to determine the total optical and camera throughput, and to reconstruct shower energies and source fluxes.

This document is structured as follows. Section 2 describes the VERITAS telescopes, giving a comprehensive overview of the main components that are used for both the standard operation of VERITAS and the calibration of the instrument. It also details the analysis techniques and workflows that are used to produce the standard science products of the telescope. The measurement of the total optical throughput and its evolution over time is presented in Sect. 3. In Sect. 4 we show the implementation of throughput corrections, which are applied to the instrument MC simulations to account for instrument aging. The impact of the varying throughput on the performance of the instrument is evaluated in Sect. 4.3, using metrics such as the energy threshold of the analysis, the differential sensitivity and the uncertainties in the reconstruction of fluxes. The validation of the method using real data collected over seven years is shown in Sect. 5. Finally, we present a brief discussion of systematic uncertainties, limitations and possible improvements of the method in Sect. 6.

2. The VERITAS telescopes

VERITAS is a ground-based very high energy (VHE) instrument operating at the basecamp of the Fred Lawrence Whipple Observatory in southern Arizona, USA (31° 40′N, 110° 57′W, 1268 m elevation). It consists of four 12 m IACTs which use a shower imaging and moment analysis technique (Weekes 1996) to detect γ-ray photons with energies above 85 GeV.

VERITAS operates in stereoscopic mode, with the array only triggering when a Cherenkov shower is detected by at least two telescopes. This allows a more accurate reconstruction of the shower properties (in particular, the incident direction and energy of the primary particle) while reducing the number of accidental triggers of a single telescope by NSB fluctuations or local muons. It also makes the discrimination of hadronic showers more efficient, which in turn boosts the sensitivity. An example of a stereoscopic reconstruction of a real EAS by VERITAS is shown in Fig. 1.

thumbnail Fig. 1.

Extensive air shower reconstructed by the VERITAS telescopes. The shower images for each camera have been integrated over time and cleaned using a two-level filter. Dead and disabled channels are shown in dark gray and brown respectively. The signal registered by each PMT is color-coded, with red colors representing higher signal than blue tones. An approximate geometrical reconstruction of the shower core location is illustrated with red dashed lines and a red star. The upper right plot shows the trace of one of PMT (#255) of T1 for reference. There, the signal registered for each sample of 2 ns is plotted in red and the integration window (six samples) is shown in shadowed blue.

The first VERITAS telescope (T1) started operations in February 2005 (Holder et al. 2006). Three more telescopes (T2, T3 and T4) were added in the following years. The full four-telescope array was commissioned by 2007 (Krennrich et al. 2007). The observatory was subsequently upgraded to an optimized diamond-shaped array layout by moving T1 to its current position in 2009 (Perkins et al. 2009), and the camera and electronics were replaced with improved hardware and higher quantum efficiency (QE) PMTs in 2011–2012 (Kieda 2013). With its current, post-upgrade, array configuration, VERITAS has been able to characterize γ-rays with energies from ∼85 GeV to > 30 TeV, and can detect a point-like source with ∼1% of the Crab Nebula flux in 25 h.

2.1. Telescope design

The VERITAS telescopes follow a Davies-Cotton design (Davies & Cotton 1957). The primary mirror of each telescope, hereafter also referred to as the “dish”, consists of 345 identical hexagonal facets. The resulting optical focal ratio of the system is f/1, for a focal length of 12 m. Each facet is designed as a one-piece glass element with ≈61 cm sides (Roache et al. 2008). They are commercially ground and slumped for a radius of curvature of 23.97 ± 0.01 m on average, and a spot size of < 10 mm. The coating is made of aluminum, deposited at a rate of 3 − 8 nm s−1 under vacuum conditions with a purity better than 99.999%. Its total thickness is 180 nm, the top 80 nm of which are oxidized during an anodizing process to improve durability and ensure a peak reflectivity of 92% at about 320 nm (Roache et al. 2008).

Cherenkov light collected by the dish is focused onto the camera. In order to detect the brief Cherenkov flashes, the cameras of the current configuration are composed of 499 high-quantum-efficiency (Otte et al. 2011; Kieda 2011) Hamamatsu R10560 PMTs located at the focal plane. Each PMT has a FoV of 0.15° and is equipped with a Winston cone, a nonimaging device designed to minimize light collection gaps between the PMT photocathodes and limit the acceptance angle (Winston 1974, 1976), therefore reducing the background light intensity. Each Winston cone is made of plastic with an evaporated aluminum coating and a protective overlayer. This provides a reflectivity greater than 85% above 260 nm (VERITAS Collaboration 2007). After a careful optical alignment of the system (McCann et al. 2010), the on-axis point spread function (PSF) of the telescopes is about 0.10 − 0.12° in diameter, leading to the ability to concentrate 80% of the light within one pixel in the camera. The optical PSF of the telescopes is frequently monitored, with observed variations of less than 0.02° over long periods of time and for most elevations. Up to 0.6° off-axis the degradation of the optical PSF with respect to on-axis observations is small, ≲0.02°, but it quickly increases at larger angles reaching ∼0.2° at 1.2° offset. The total FoV of the telescopes is 3.5°.

2.2. Readout and trigger system

PMT signals are digitized using 500 Msample s−1 flash analog-to-digital converters (FADCs) with 8-bit dynamic range, capable of storing the waveforms in 64 μs memory buffers. VERITAS employs a three-level trigger. A first-level trigger (L1) or constant fraction discriminator (CFD) requires the PMT pulse height to be above a given threshold (typically 5−6 photo-electrons). A pattern trigger or telescope trigger (L2), requires L1 trigger signals to occur in three adjacent PMTs within the timing coincidence window. This pattern trigger is based on 400 MHz Xilinx Virtex-5 FPGAs for pixel neighbor coincidence, allowing triggers to occur before samples are read out. The time-aligning accuracy of this system is ∼0.2 ns and allows for a pixel-to-pixel coincidence window of ∼5 ns. Finally, an array trigger (L3) requires a L2 trigger signal from at least two telescopes to happen, once corrected for the propagation time of the shower front across the array and the varying distance from each telescope to the central control building, within a 50 ns coincidence window. A more detailed description of the VERITAS trigger system can be found in Zitzer (2013).

Coincidence window widths and CFD thresholds are optimized to trigger on low-energy gamma-ray showers while avoiding random coincident triggers from NSB. The optimization process consists of scans of the trigger thresholds when the instrument is exposed to NSB. The aim is to keep the L3 rates at a few hundred Hz, and to balance a low trigger energy threshold with avoiding data losses from dead-time. The typical dead-time for VERITAS, in its current configuration, is roughly 15% for a data acquisition rate of about 300 Hz.

2.3. Data analysis

VERITAS maintains two data analysis packages: Eventdisplay (Maier & Holder 2017) and VEGAS (Cogan 2008). They allow independent reconstruction of the data, limiting the impact of systematic uncertainties due to the analysis software implementation on the scientific results. Each package performs a calibration of the signal collected for each shower by the four telescopes, a second-order moment analysis, and parametrization of the shower images. The resulting parameters are used to classify the showers as γ-ray-like showers or hadronic events. Using the stereo shower reconstruction, the arrival direction and the energy of the primary particle are estimated. Finally, from a comparison with reconstructed MC shower simulations, the effective collection area is evaluated and the events in excess of the estimated background are converted into high-level analysis products such as fluxes, light curves, and energy spectra. The results of the throughput analysis shown in this work were obtained using the Eventdisplay package, however they were also validated using VEGAS.

2.3.1. Calibration

The calibration process comprises the determination of the electronics baseline (pedestal) and its variation plus the measurement of the response of each individual PMT to incident light, that is absolute and relative gain differences between PMTs. FADC inputs are AC coupled and a DC voltage offset is added to the signal inputs (pedestal) so that positive and negative fluctuations around the mean value (pedestal variation) due to NSB variations can be measured. Artificially triggered pedestal events are injected during observation runs with a frequency of 1 Hz and used for the estimation of pedestal level and variation. The sky brightness and therefore the background light level might change during observations; pedestal levels and variations are consequently updated every three minutes.

PMT gains are monitored and calibrated by using a flasher light source. The VERITAS FADCs have two gain channels for a wide dynamic range of the readout chain (Holder 2005). The calibration software reconstructs the relative gain of the pixels, timing differences (e.g., due to differences in the cable length for each channel), and relative calibration of the high voltage settings and the high- and low-gain readout channels by using uniformly illuminated camera events generated with flasher light sources (Hanna et al. 2010). Each flasher unit consist of blue (370 nm) light-emitting diodes (LEDs), driver electronics, and a front face made of an opal diffuser which spreads the light from the LEDs and distributes it almost homogeneously across the entire PMT camera. The flasher pulses span eight brightness levels (upgraded to fifteen in 2016), covering the dynamic range of VERITAS for both high- and low-gain readout. Absolute gain calibration is determined on a periodic basis following the procedure described in Sect. 3.1. Relative gain differences between PMTs are monitored daily, and corrected during the data analysis. The inter-calibration between the two gain channels is performed on a monthly basis by recording calibration runs with a particularly long readout window. This is needed to avoid truncation and provide samples at the end of the trace that allow us to estimate the low-gain pedestal. For each run, half the camera is operated at a reduced gain to force it to stay in high-gain mode, while the other half operates in low-gain.

Finally, the results of the analysis of Cherenkov light from muons measured by the VERITAS telescopes (Tyler 2013) are used to monitor the calibration, in conjunction with the procedures described below (single photo-electron runs, optical PSF, and mirror reflectivity measurements). This approach mitigates several limitations of a calibration based on muons only, especially the differences in the wavelength spectrum between Cherenkov light from muons and from air shower events.

2.3.2. Signal extraction and image analysis

Signals are extracted from the FADC traces using a two-pass integration method. The first pass uses a long integration window to determine the pulse arrival time and to derive the optimal position of the time integration window, using a linear fit of the pulse arrival times along the image axis (Holder et al. 2006). A second pass performs the trace integration, using a short window of typically six samples (12 ns). An example of a PMT trace for a real event and the optimized integration window is shown in Fig. 1.

A two-level filter is then used to extract the signal pixels and clean the image. First, the brightest parts of the shower image (core pixels) are localized at five times the pedestal root mean square (rms). If a Gaussian distribution of pixel signals is assumed for images containing only noise, this would correspond to a surviving rate of less than one in 3 × 106. This is enough to remove most clusters of pixels containing only noise that might exist in the full image. Then, to avoid cutting out the edges of the shower images, adjacent (boundary) pixels with at least 2.5 times the pedestal rms are added to the core pixels. The resulting shower image, which resembles an ellipse in the case of γ-rays, is parameterized using a second moment analysis, based on the method originally developed by Hillas (1985).

Images from showers with large inclinations, large offsets, or originating from high energy showers may not be fully contained in the camera. These “leaking events” cause a loss of signal, degrade energy resolution and take a toll on the performance. Using a log-likelihood fitting algorithm, some of these showers are recovered (Maier & Holder 2017), particularly improving the performance of the array at high energies.

2.3.3. Direction, shower core, and energy reconstruction

The direction of the primary particle and the impact parameter of the shower core on the ground are derived using stereoscopic techniques (Hofmann et al. 1999; Maier & Holder 2017). The energy of each γ-ray event is estimated from a comparison using lookup tables built with simulations. The lookup tables encode the dependency of the distance of the air shower core to each telescope, integrated signal of the shower image (size), noise level, azimuth, zenith angle and array configuration.

2.3.4. Gamma-hadron separation

Cosmic rays represent the largest fraction of the data for all γ-ray observations1. In order to separate signal from background cosmic-ray events, it is common practice to compare the shape of observed shower images with simulated γ-ray showers. To break the dependency of length and width of the image ellipse on the energy of each event, Eventdisplay uses the median reduced scaled parameters (MSCP) of width (MSCW) and length (MSCL; Krawczynski et al. 2006):

MSCP = 1 N tel i = 1 N tel p i p ¯ MC ( θ , ϕ , α , NSB , ATM , size , r ) σ 90 , $$ \begin{aligned} \mathrm{MSCP} = \frac{1}{N_{\rm tel}} \sum _{i=1}^{N_{\rm tel}}{\frac{p_i - {\bar{p}}_{\rm MC} (\theta , \phi , \alpha , \mathrm{NSB}, \mathrm{ATM}, \mathrm{size}, r)}{\sigma _{90}}}, \end{aligned} $$(1)

where pi is the measured parameter (in our case width or length) for telescope i, p ¯ MC $ \bar p_{\mathrm{MC}} $ and σ90 the expected value and width of the distribution for 90% of the events, which are read from lookup tables that are filled using simulated gamma rays. Finally, θ is the zenith angle, ϕ the azimuth angle, α the offset angle of the telescope pointing direction with respect to the nominal source position, NSB the night sky background, ATM the atmospheric profile, size the integrated charge of the shower image, and r the distance of the shower core to telescope i. VEGAS, in contrast, uses nonreduced versions of these parameters, centered at 1 instead of 0.

The separation of γ-rays from background cosmic-ray events is done using boosted decision trees (BDTs; Krause et al. 2017). BDTs are implemented as part of the Toolkit for Multivariate Data Analysis (TMVA) package (Speckmayer et al. 2010) and trained on a set of simulated γ-ray and observed hadronic showers using reconstructed shower parameters: MSCW, MSCL, shower core position, height of the maximum of the Cherenkov emission, and quality of the energy reconstruction.

With different science goals in mind, three sets of cuts are usually defined in VERITAS, each requiring different BDT scores and different minimum image brightness. “Soft cuts” provide a low energy threshold but larger acceptance of background events. They are ideal for sources with very steep energy spectra which emit large fluxes from tens of GeV to a few hundreds of GeV. “Moderate cuts” are the most widely used and are adequate for most analyses, as they provide a good balance between energy threshold and sensitivity. “Hard cuts”, particularly combined with higher multiplicity (e.g., 3-telescopes) provide the best background rejection. Though the energy threshold is high (≳300 GeV), this set of cuts is used when the source of interest is weak but with emission extending well into the TeV energy range. Unless otherwise stated, “moderate cuts” have been used throughout this work.

2.3.5. Spatial and spectral reconstruction

Event counts are converted into energy spectra taking into account the effective areas of the telescopes and dead-time corrected exposure time (Acciari et al. 2008). The background at each point in the sky is estimated using either the reflected region method or the ring-background model approach (Berge et al. 2007).

2.4. Monte Carlo simulation

Instrument response functions (IRFs), required for both the separation of γ-rays from background events and for the reconstruction of energy spectra and light curves, are generated using large sets of simulations of γ-ray showers.

The propagation of extensive air showers in the atmosphere is carried out using the CORSIKA package (Heck et al. 1998) taking into account all relevant particle interactions. The generation, propagation, scattering, and absorption of Cherenkov light is simulated using measurements of molecular and aerosol content of the atmosphere from a radiosonde (station number #72274) which operates from Tucson, Arizona, approximately 60 km away from the VERITAS site. From the radiosonde data, two sets of simulations are generated, the first with an average winter atmospheric profile and the second for summer atmosphere.

The response of the instrument is simulated using the GrOptics2 and CARE3 packages for the optical and camera simulations. The optical simulations take into account the properties of the primary mirrors including facet alignment, reflectivity, diffuse scattering and all relevant shadowing elements of the telescopes. The photo-detector and electronic simulation code CARE is used to model the digitization, trigger, and readout processes including noise components. The MC model is validated through extensive comparisons with calibration measurements.

The generation of MC simulations for instrument response functions is an important computational effort, given the large parameter space that needs to be covered. The MC instrument model for the array configuration after the upgrade of the cameras was generated to accurately reproduce the array performance for the period October 2012–June 2013.

3. Throughput measurements

The collection efficiency of Cherenkov light at ground level by the telescopes and its conversion to measurable signals, that is telescope throughput, strongly depends on the properties of the cameras and dishes, which change over time. At the camera, PMT gains and the QE affect the conversion of photons into charge pulses in each pixel. Similarly, the combined effect of the reflectivity of the mirror facets and the collection efficiency of the Winston cones affects the total optical light collected at each PMT photocatode. Furthermore, some of these parameters vary on different time scales and some are wavelength dependent, turning the characterization of the different parts of the instrument into an important and very challenging step in the analysis of data produced by IACTs. This section describes how these components are studied in VERITAS.

3.1. Camera gains and gain factors

The current detector model of VERITAS assumes an average value of the absolute PMT gain Gi, MC for each telescope i, estimated after the upgrade of the PMT cameras in 2011–2012:

  • G1, MC = 5.38 dc/pe

  • G2, MC = 5.29 dc/pe

  • G3, MC = 5.29 dc/pe

  • G4, MC = 5.73 dc/pe

where dc stands for digital counts and pe for photo-electrons.

In practice, the individual absolute gains are slightly different for every PMT and they vary over time due to temperature fluctuations. Supply voltage fluctuations are are kept stable at the subvolt scale, therefore are unlikely to have a large impact in the absolute gains.

At the beginning of each observing campaign, usually in October, absolute and relative gains are measured and adjusted by changing the HV settings of the telescopes. The process is usually repeated once or twice during the season. This brings the gains closer to the nominal values shown before and results in absolute gain swings of, at most, 10% over months. The resulting PMT charge distribution is also optimized to have ≲5% width. Part of the aim of this work is to reduce the impact on the IRFs of the remaining gain variations left after the flat fielding process.

Gains are measured in VERITAS in two independent ways, each with its own set of assumptions. The first approach directly measures the absolute gain by detecting the signal from single photo-electrons. This is achieved by placing a cover to reduce the optical transmission in front of the camera. The second method evaluates the absolute gains of the PMTs statistically, using “LED flashers” to uniformly illuminate the camera. Flasher runs are collected daily to perform a relative gain correction directly in the analysis chain.

3.1.1. Single photo-electron (SPE) runs

Weak, pulsating sources of light can be used to study the response of PMTs to a single photo-electron. The measured charge from a single photo-electron is, on average, proportional to the number of electrons produced by the PMT. The proportionality constant is the gain of the system, G (see e.g., Kieda 2011; Hanna et al. 2010; MacLeod 2007).

In VERITAS, single photo-electron light levels are achieved by attenuating the light of the flasher with a custom-made camera cover with a small hole aligned with the center of each PMT (Hanna 2008). Each of these holes allows only about 1.6% of the light to pass. With such a device, not only are the required low-illumination conditions achieved, but NSB is suppressed by the same amount.

A histogram of the accumulated charge shows a series of peaks. The first peak describes the pedestal, the second the mean value of the SPE charge in digital counts. The separation depends on the absolute gain set by the HV settings used during the data acquisition, making it possible to derive the gain-voltage relationship and measure the absolute gain. Since the absolute gain can locally be approximated as having a linear dependence on voltage, VERITAS takes SPE runs at both nominal and a slightly increased voltage (110%) so that the two peaks are better separated.

The shape of the single-electron peak follows a Polya distribution (López Paredes et al. 2018), which is a special case of a negative-binomial distribution. The comparison to photo-statistic gains, described in the next section, requires a correction factor for each pixel, which quantifies the deviation from Poisson statistics. SPE runs are taken in VERITAS approximately once a month, as they require the manual installation of the custom cover on each of the telescope cameras, temporarily interrupting the data taking and leading to increased observer fatigue and safety risks.

3.1.2. Photo-statistic gains from nightly flasher runs

The second method to measure absolute gains in VERITAS uses nightly flasher runs. We refer to these gains as photo-statistic gains. During a flasher run, the mean charge of a pulse on a PMT, μ, is statistically proportional to the mean number of photo-electrons at the first dynode, Npe:

μ = G × N pe . $$ \begin{aligned} \mu = G \times N_{\rm pe}. \end{aligned} $$(2)

The proportionality constant is the absolute gain G. It can be determined by assuming that Npe follows Poisson statistics, which implies that the variance in photo-electrons is approximately the mean charge μ, hence σ pe N pe $ \sigma_{\mathrm{pe}} \simeq \sqrt{N_{\mathrm{pe}}} $. After the dynodes and preamplifier have amplified the signal, the variance becomes

σ 2 G 2 × N pe = G × μ . $$ \begin{aligned} \sigma ^2 \simeq G^2 \times N_{\rm pe} = G \times \mu . \end{aligned} $$(3)

Fitting a linear function to the charge variance σ2 over μ provides the gain as the slope and the variance of the pedestal noise as the intercept.

3.1.3. Comparison of absolute gain measurement methods

The two methods of measuring the gain agree within 5 − 6%, which is assumed to be the systematic uncertainty for the absolute gain estimation. Photo-statistic gains can be determined using runs that are taken once a night, therefore allowing better monitoring. We consequently opted to use them for our throughput calibration purposes, as opposed to SPE gains.

So far we have only discussed the evolution of the gains, but the QE of the PMTs may also change as the detector ages. The possible aging of VERITAS PMTs was covered in Gazda et al. (2016). In that work, a set of 20 PMTs were removed from the telescopes and compared to 20 spare PMTs that were never operated and therefore are expected to exhibit their original properties. The absolute quantum efficiency was measured to be consistent for both samples at the wavelength range in which VERITAS detects Cherenkov radiation from γ-ray showers. More recently, during the Summer break of 2020, another aging study was done using the same sets of PMTs. This time, the test consisted of the determination of the HV-dependence of PMT gains, which is related to their QE. Again the properties of both sets of PMTs were found to be compatible within uncertainties.

3.1.4. Gain factors

Once the average absolute camera gains (Gi) are measured for each telescope i, they can be compared with the gains assumed in the reference MC simulations (Gi, MC) to obtain the gain factors (gi = Gi/Gi, MC) needed to correct the simulated signals to account for changes in the camera gains. Figure 2 shows the gain factors as derived from photostatic gains.

thumbnail Fig. 2.

Time-dependent changes of photo-statistic gains normalized to the nominal absolute gains used in the baseline MC model for the current detector model. This detector model was constructed in 2012, and reproduces the characteristics of VERITAS after the PMT camera upgrade. Black points show the average gain factors for each instrument epoch. Gray points show the individual unfiltered gain factor values per telescope and daily flasher run. Blue, orange, green and red points show the result of a median filter of the individual gain factors. Curves of the same colors represent a spline interpolation of the filtered values. Even though spline interpolation can locally diverge when there is no data available, we note that this occurs only during Summer breaks, where no data is taken in VERITAS, therefore it is not a concern for this study. Error bars represent statistical uncertainties.

3.2. Optical throughput and mirror reflectivity

VERITAS is located in southern Arizona, where at least three months during summer are under direct influence of the North American monsoon, in addition to the all year round varying weather conditions. The mirror facets of the telescopes are consequently affected by mechanical and chemical degradation of their reflecting surfaces (e.g., due to fine dust grains which scratch the reflective coating, pollutants chemically reacting with the mirror materials, or temperature oscillations inducing deformations on the underlying mirror substrate). Furthermore, periodic mirror cleaning, re-coating, and Winston cone cleaning contribute to partially recover the reflectivity of VERITAS telescopes, but it also adds an additional variability component to the optical performance of the telescopes.

The degradation of the mirrors changes their reflective properties. The total reflectivity is the combination of diffuse and specular reflectivity. The first is the tendency of a surface to reflect the incident radiation in random directions. Degraded mirror surfaces can cause high diffuse reflectivity, scattering photons and making the mirrors unable to focus and properly form an image. For imaging instruments, diffuse reflectivity is undesirable as it increases background photon noise at the focal plane and degrades the optical PSF. Specular reflectivity refers to the ability to form an image of a given object at the focal plane. It is very sensitive to mirror aging since it depends on the material of the substrate, how accurately the mirror is figured, how smooth its surface is and the type of coating.

Mirror coating degradation causes a wavelength-dependent effect on the optical throughput of the instrument (Garoli et al. 2020), with reflectivity at short-wavelengths more severely affected than that for longer wavelengths. This is relevant for IACTs because the bulk of the Cherenkov emission concentrates in the near-UV and blue part of the visible spectrum.

The optical properties of the Winston cones are not expected to vary dramatically over time as they are protected from the elements and the camera is fully closed while the telescopes are stowed (i.e., during daylight and during bad weather conditions). Nonetheless, these components were examined over time and no evidence of degradation in their light collection efficiency was detected. Therefore, we concentrate on the reflectivity of the telescope dishes and treat Winston cone efficiency as a source of systematic uncertainty.

3.2.1. Laboratory measurements of specular reflectivity

Beginning with the VERITAS inauguration in 2007 and continuing until 2015, fifteen facets from different parts of the dish (top, middle, bottom) were regularly removed to monitor the changes of the wavelength-dependent reflectivity over time. This was done in the laboratory using a broad-spectrum light source, an adjustable filter wheel, and a photometer (Oriel 71610). Measurements were compared to a calibration mirror made of pure aluminum, periodically re-coated to ensure consistency in the measurements (Roache et al. 2007).

Even though it was intented to measure specular reflectivity, this method had the disadvantage of not being able to fully discriminate between diffuse and specular reflectivity as it relied on a photometer and not on a real imaging detector. It also did not evaluate the impact of PSF changes on the performance of the instrument collecting Cherenkov light. Finally, the evolution of the reflectivity of a few mirrors was measured, rather than that for the entire dish.

Laboratory measurements of specular reflectivity were dropped in 2016 as the achieved accuracy was limited by the complexity of determining the ratio of specular to diffuse reflectivity.

3.2.2. Whole dish reflectivity (WDR)

The WDR method was initially developed by MAGIC (Mirzoyan et al. 2007) and after a few preliminary tests in 2010 and 2011, the method was adopted by the VERITAS observatory starting from 2014 (Archambault et al. 2013, 2021). The WDR technique uses a wide-field CCD camera attached to the dish, near its center, which is pointed directly toward a target at the focal plane in front of the PMT camera. In order to reproduce the response of the mirrors to Cherenkov light combined with the QE of the PMT, which peaks at ∼340 nm, the CCD (SBIG ST402ME) is equipped with a blue filter from the same manufacturer (Diffraction Limited©). This effectively shifts the response of the CCD camera to shorter wavelengths and provides a better match to the spectral response of the PMT camera to extended air showers. The target is made of Spectralon, a fluoropolymer which exhibits Lambertian behavior (i.e., high diffuse reflectivity). This ensures that the apparent brightness is the same regardless of the observing angle. The reflectivity of the Spectralon plates used in VERITAS has been checked over the years and no significant change could be detected, setting a conservative limit of ≲5% to their evolution, well within the total systematic uncertainty of the reflectivity determination. With this set-up, the CCD camera can be used to measure simultaneously the brightness of a reference star in the FoV and its reflection generated by the dish on the Spectralon. Flat and dark frames are used to correct for effects such as vignetting or pixel-to-pixel response, and to subtract the electronic dark signal in the image. The comparison of the brightness of the reference star and its reflected spot yields an estimation of the total reflectivity of the dish. The advantage of this method resides on its simplicity and robustness, using just a standard CCD with a color filter and target plate over the focal plane. Being based on differential photometry methods, it is insensitive to changes in the QE or the gain of the CCD itself. The main disadvantage of the WDR method is that the employed CCD is blind to UV radiation and therefore the spectral averaged reflectivity does not fully match the reflectivity for Cherenkov light.

3.2.3. Optical throughput and optical throughput factors

In a manner similar to the case of the gains, this study focuses on the relative changes of the reflectivity with respect to what is simulated for each telescope i in the detector model. From the two possible methods to estimate the reflectivity of VERITAS mirrors, we adopted the WDR method described in Sect. 3.2.2.

VERITAS simulations take into account the complete spectral response of each telescope as a function of wavelength (see Appendix A). On the other hand, the WDR method only provides an averaged reflectivity for each telescope i, measured in the blue part of the visible spectrum. The calibration of the optical throughput requires the estimation of the same wavelength-averaged reflectivity for the reflector model of the current instrument configuration Ri, MC and the assumption that the differences due to the spectral mismatch that exists between the Cherenkov spectrum and the assumed blue filter are small.

The spectral average is determined assuming a standard Bessel-B filter Bessell (1990), similar to the one used with our CCD. The transmission of this filter is multiplied by the reference reflectivity curves from the reflector model. The resulting curve is integrated to obtain the average reflectivity Ri, MC for each telescope:

  • R1, MC = 0.828

  • R2, MC = 0.841

  • R3, MC = 0.852

  • R4, MC = 0.846

In order to estimate the collection efficiency error due to the use of a B-filter and a CCD to characterize the response of the telescope to Cherenkov light, we computed the wavelength-averaged reflectivities, this time convolving each mirror reflectivity curve with a typical Cherenkov spectrum from a γ-ray induced shower. The resulting average reflectivities are between 0.79 and 0.82 for all telescopes. The differences are therefore well within the assumed systematic uncertainty on the mirror reflectivity (≲10%).

Calculated based on the reflectivity measurements, the optical throughput factors or reflectivity factors can be defined as ri = Ri/Ri, MC. Their evolution over time is shown in Fig. 3. These reflectivity factors are needed to correct our MC simulations so that they reflect the changes in the mirror reflectivity. The first point is fixed at unity as it represents the reflectivity values actually used in the simulations. Because the whole-dish reflectivity only started to be measured routinely in 2014, the second and third points of each panel of Fig. 3 had to be interpolated. We artificially assigned them a comparatively larger uncertainty of ±5% in absolute reflectivity (±0.05 in the plot). As can be seen from the figure, reflectivity quickly degraded by 20 − 30% from 2013 to 2015, coinciding with a period of reduced cadence of re-coating degraded mirror facets. Since 2015, the telescope reflectivity continues to decrease, but at a much slower pace. Improvements in the reflectivity are also present in our data for specific times, coinciding with cleaning and re-coating of the most damaged mirror facets. This process usually involves the exchange of ∼100 mirror facets, and its effects are smoothed in the spline fitting case.

thumbnail Fig. 3.

Time-dependent changes of the reflectivity factors obtained from WDR measurements. Black points show the average reflectivity factors for each instrument epoch. Blue, orange, green, and red points show the individual reflectivity factor measurements for each telescope. Curves of the same colors represent a spline interpolation, which removes outliers. The first point (early 2012) is extracted from the simulations and it serves as the reference for the calculation of the reflectivity factors, therefore adopting a value of 1 as reflectivity factor. We note that a value of the reflectivity factor slightly greater than 1 just means that the reflectivity at that time was slightly higher than at the time where simulations were produced.

3.3. Total throughput

The total throughput of each telescope, normalized to the reference values in the simulations, can be computed as the product of the gain factor gi (Fig. 2) and the corresponding reflectivity factor ri (Fig. 3). The resulting values and their evolution for each telescope are shown in Fig. 4 and Table 1. We refer to this total throughput factor as ti or simply throughput factor.

thumbnail Fig. 4.

Time-dependent change of the throughput factors calculated from photo-statistic gain factors and WDR reflectivity factors. Error bars show only statistical uncertainties.

Table 1.

IRF periods for the current VERITAS instrument configuration.

To ensure consistency, we computed the average throughput factors for each time bin with three different strategies: (i) using the product of spline interpolations for gain factors and reflectivity factors and then computing the mean for each season, (ii) using the closest reflectivity and gain measurement pairs to compute the total throughput and then estimating seasonal averages, and (iii) using the average values of gain factors and reflectivity factors per instrument epoch and calculating their product. They all agree at the ∼5% level.

The application of these throughput factors to the Cherenkov signals of the reference simulations allows us to produce IRFs that account for the throughput degradation of the VERITAS telescopes. The actual implementation and assessment of this calibration is discussed in the next section.

3.4. IRF period definition

VERITAS has undergone two major instrument upgrades. The first consisted on the relocation of T1 to its current position, improving significantly the trigger efficiency. The second upgrade targeted the camera, with improvements to the electronics and the PMT QE. This naturally generated three major instrument configurations, each with its own set of MC simulations, which provide an accurate model of the telescope performance.

The current instrument configuration spans almost eight years of operations, during which the instrument has evolved significantly due to aging and our maintenance efforts. Consequently, we tune our IRFs and simulations in finer time bins so that each time bin or IRF period has throughput drops of at most 10%, comparable to the claimed systematic uncertainties on dish reflectivity and gains. Similarly, we require the throughput values at the limits of each bin to be covered by the statistical errors for that bin. Finally, the duration of each epoch is at most a single year, and approximately aligned with the observing cycle that begins with the end of the monsoon in September or October.

With these criteria it is clear that for the last years, during which the evolution of the throughput was less dramatic, one IRF period per year is enough to provide the desired granularity and precision. For the first couple of years, however, the total throughput was rapidly evolving and it required finer time bins to fulfill our criteria. The resulting IRF periods are summarized in Table 1.

4. Implementation of throughput calibration

This section discusses various strategies to implement the throughput calibration in IACTs, a description of the method adopted in the VERITAS software packages and the tests we performed to ensure the consistency of the results.

4.1. Implementation methods

MC simulations should accurately reproduce the response of VERITAS at any given time. We discussed in the previous section the impact of the time-dependent changes on the PMT gains and optical elements of the telescope. Reproducing all the details with the required accuracy while using a single MC production and a static set of IRFs is not an option if throughput is changing over time. Here, we concentrate on how we can implement the throughput calibration in our analysis framework to ensure consistent detector response over the entire operating life of the telescopes.

There are a number of strategies one could adopt to implement this calibration in the analysis, with varying degrees of accuracy that correlate with the computational cost and the parameter space. In the case of VERITAS, we need to cover the response of the instrument as a function of zenith angle of observation, azimuth angle, atmospheric profiles, telescope multiplicity, camera HV settings, wobble pointing offset, and NSB intensity.

Full MC productions that reproduce the changes in the throughput of the instrument over time and corrections to an existing MC production are both valid possibilities. Provided that the systematic errors on the calibration measurements are small, the former is potentially the most accurate and correctly reproduces the changes at the trigger level caused by throughput variations. However, it is computationally very expensive. It also requires producing a new complete MC model of the entire array and simulating γ-ray showers with the required statistics. Consequently we chose to explore corrections to the already existing MC simulations that were produced for the baseline instrument configuration.

Calibration corrections could be applied on a run-wise basis (with typical duration of an observing run shorter than one hour) or on longer periods of time: from just one day to an entire year or even longer, during which the instrument response is measured to be approximately stable. Because of the monsoon shutdown in summer, which defines a natural observing season from about mid-September to mid-June, the fact that camera voltages (and hence gains) are adjusted at the beginning of each of these seasons and the slow cadence of WDR measurements, such seasons seemed a natural choice for VERITAS to define the time bins for the corrections. Exceptionally for the first two years, for which the throughput evolved rapidly, each season was divided into two shorter time bins. We refer to these time bins in which we apply the throughput corrections as IRF periods.

Throughput corrections can be introduced at different steps in the analysis of simulated showers and the production of the IRFs. A simple option, yet not fully accurate, is to calibrate simulated event images after they have been cleaned and the PMT traces have been integrated. The size of the event is modified using the changes in the total throughput of the instrument. Noise level changes can be either ignored or approximately corrected. Testing this option in VERITAS resulted in a reasonable reconstruction of event energies and γ-ray fluxes. However, the corrected simulated image shapes did not match the shape of the observed shower images.

A natural improvement over this, and the solution adopted in this study, consists of correcting the signals of the simulated events before the PMT traces are integrated and the Cherenkov images are cleaned. Even though the trigger changes are still not properly modeled, noise levels change as fewer pixels containing only NSB are able to pass the image cleaning thresholds. Our tests resulted in reconstructed simulated showers which are directly comparable to observed data, making it easier to assess the accuracy of the corrections by comparing the distributions for the different image parameters. The main disadvantage is that the cleaning and Hillas parameterization of the simulated showers had to be repeated once per IRF period with each throughput factor ti on the signal integration, significantly increasing the computing time, storage requirements and analysis complexity needed to cover the entire parameter space. Producing IRFs for a single wobble offset with 10 time bins as described in this work, including two atmospheric profiles, and all zenith angles (8) and noise levels required (11), took several weeks of testing and computation time and required about 480 GB of storage for the final products.

4.2. Robustness of shower parameter reconstruction

This work is based on the assumption that we can correct the simulated PMT traces with just two global factors, one due to the camera gain gi and the other due to the dish reflectivity ri. In principle, this should not distort the shape of the shower images; instead we expect it to alter the size of the cleaned events. To check if this is the case, we extracted the same simulated events corresponding to two different IRF periods (2012–2013a and 2019–2020) and calculated the ratio of values for some of their key geometrical parameters (size, width, length, and time gradient along the major axis) between the two periods. Figure 5 shows histograms of the obtained ratios together with the mean and the standard deviation of the obtained distribution. It can be seen that only the size of the events is significantly shifted (μ ≃ 0.63, σ ≃ 0.34) when we scale the simulated pixel traces. As this is a statistical comparison between individual simulated showers obtained with independently generated IRFs, the large width of the distributions for some of these ratios is not a concern for this study.

thumbnail Fig. 5.

Ratio between image parameters derived from MC simulations for the season 2019–2020 vs season 2012–2013a (Telescope T1 only, winter simulations, noise level of 100 MHz, and no size cuts). The vertical axis of each panel shows the number of events within a given ratio bin.

4.3. Impact of throughput changes on VERITAS performance

The evolution of the total throughput can affect the reconstructed fluxes in two different ways. The relation between the size of an event and its energy changes as discussed in Sect. 4.3.1. Additionally, some events are no longer reconstructed, as their integrated charges are not large enough to pass the hardware trigger, analysis, and reconstruction cuts. This has a significant impact for the lowest energies, as is shown in the effective areas of the telescopes, discussed in Sect. 4.3.2.

4.3.1. Effect on reconstructed energies

One of the main effects of the application of correction factors to the measured digital traces is the improvement of the energy reconstruction. Lowered gain and reflectivity naturally result in smaller images and consequently lower estimated energy if the IRFs do not account for throughput variations. The energy reconstruction method applied uses lookup tables, similar to the ones described in Sect. 2.3.4. They represent the energy as a function of size and distance to the telescope in a coordinate system perpendicular to the arrival direction of the air shower event, the NSB, the wobble offset of the observation and the atmospheric profile. These tables are filled using simulated γ-ray events by calculating the median and the standard deviation of the distribution of true energies (EMC) for each bin in the grid.

Figure 6 shows the ratio between the reconstructed energies of the events whose traces have been corrected to account for the drop in telescope throughput and the uncorrected case. That ratio is represented, following the format of the lookup tables, as a function of impact distance and size of the simulated event for a fixed set of values for the other parameters.

thumbnail Fig. 6.

Relative changes in energy reconstruction induced by the evolution of the throughput of the instrument. The ratio of throughput-calibrated energies with respect to the uncorrected energies is presented in two formats: As 2D lookup tables, with the ratio as function of impact distance and size of the simulated events (bottom panels); curves with the distance-averaged lookup-tables, showing the ratio as a function only of size (top panels).

For a given noise level and zenith angle, the size of the event scales almost linearly with its energy. The exception would be the highest energies. In that regime, the accumulated charge may hit a maximum value due to FADC saturation, affecting the linear relation between size and energy previously described. Since the events migrate to lower sizes for a fixed energy as the telescope throughput drops, one can easily see the effect of its evolution with time once we calibrate the throughput in the analysis: for a given image size and distance of the event, the reconstructed energy is higher in recent years than it was right after the upgrade of VERITAS (years 2012–2013).

4.3.2. Effect on flux reconstruction and effective areas

The impact of the total throughput on the effective area is shown in Fig. 7. In that figure, the effective area is computed as a function of true energy in each period for a fixed position in the parameter space (zenith angle of 20°, noise level of 100 MHz, moderate cuts, a minimum of two images per event, and 0.5° wobble offset).

thumbnail Fig. 7.

Evolution of the effective area for different instrument epochs, for a zenith angle of observation of 20°, noise level of 100 MHz, moderate cuts and a minimum of two images per event.

There is insufficient background data to optimize the cuts individually for every period. The reason is that we require data that were taken under very good weather conditions, fields without strong γ-ray sources, and dividing the remaining data sample into zenith angle bins. Consequently, we used the same set of cuts optimized using data from the entire post-upgrade array configuration. This artificially enhances the differences seen at low energies in Fig. 7, since fewer events have the size required to pass the analysis cuts in the latest periods. Reducing the size cut over time to palliate this effect is not an option as it would result in the inclusion of worse quality events in the analysis. Above ∼300 GeV, away from the energy threshold of the analysis, almost no events are lost despite the decreased optical throughput of the telescope.

4.3.3. Effect on the energy threshold of the analysis

Due to the statistical nature of the reconstruction methods used for IACTs, the energy threshold of the analysis is defined as the energy at which the effective collection area is 10% of the maximum effective area reached with the telescope at any energy. This definition ensures that a significant fraction of the events have passed the cuts and that the individual shower images and traces are comparatively less affected by noise artifacts. It does not imply that lower energy particles are undetectable, but the percentage of particle showers below this energy quickly drops as the energy decreases. The detected emission at energies lower than the energy threshold may still be significant for sources with steep spectra.

Figure 8 shows the evolution of the energy threshold of the analysis with VERITAS. Uncertainties of the energy thresholds shown in the figure are calculated accounting for two independent contributions:

  • (i)

    the energy bin width of the effective area histogram,

  • (ii)

    the uncertainty on the effective area at the center of the bin.

thumbnail Fig. 8.

Energy threshold of the analysis estimated from simulated events and using moderate cuts, zenith angle of 20°, and noise level of 100 MHz. It is defined as the energy for which the effective area becomes 10% of the maximum effective area.

4.3.4. Effect on the differential sensitivity

We can define the sensitivity of a γ-ray instrument as the minimum flux that such an instrument can detect for a given amount of exposure. Simulating a realistic sample of background data is challenging, so it is common practice to use the Crab Nebula data as a reference object and to use its flux as the unit system for sensitivity studies. In the following, the differential sensitivity is computed as the minimum photon flux required to achieve a 5σ detection in 50 h of observation time (using the statistics described in Li & Ma 1983) in each energy bin, assuming five equal-width bins per decade on the logarithmic energy axis. Additionally, at least ten signal events are required per energy bin.

The differential sensitivity is presented in Fig. 9, based on good quality Crab Nebula data, which were collected under dark sky conditions, and at low zenith angle to minimize the atmosphere’s influence. We show the results for BDT-based moderate cuts with a minimum telescope multiplicity of two, which provides a balance between optimizing sensitivity and maintaining a low energy threshold. Results employing softer cuts and harder cuts were obtained and compared between seasons, yielding similar conclusions.

thumbnail Fig. 9.

Differential sensitivity with 50 h of VERITAS observations calculated using different low zenith angle Crab Nebula data, with winter atmospheric profile conditions. The same cuts as in Fig. 7 are used. The sensitivity is defined in terms of the flux of the Crab Nebula (in Crab units, CU) assuming it is a reference astrophysical object at VHE energies.

It can be concluded that throughput degradation affects mostly the lowest energies, where some dim shower events are lost or mistakenly classified as hadronic showers. At high energies, the performance is similar between the different periods, unaffected by the changes in the throughput.

5. Validation of the throughput calibration on real data

5.1. Comparison between real data and MC simulations

The validity of the estimated total throughput correction can be tested by comparing image and stereo parameter distributions obtained with data and throughput-calibrated simulated events for every IRF period. If MC showers are not calibrated with the correct throughput, the shower shapes will be distorted and the estimated shower parameter distributions will not match those derived from data.

In this work, we compared MSCW and MSCL distributions of simulated and observed data for six reconstructed energy bins: log10Erec ∈ [ − 1.0, −0.7], [ − 0.7, −0.3], [ − 0.3, 0.0], [0.0, 0.3], [0.3, 0.7], [0.7, 1.0] (energy measured in TeV), and for all the considered IRF periods.

The results, with the last two energy bins combined to increase statistics, can be seen in Fig. 10. More detailed comparisons are shown in Appendix B. In all cases, data and simulated events have similar distributions for both parameters, almost symmetric, with similar widths and centered at zero (as expected for γ-ray events). No significant evolution was observed in these parameters with time. The only exception is the highest energy bin, which has a wider distribution of MSCW values for real shower events. This is caused by the high energy events, ∼10 TeV, being subject to difficulties in calibrating of the low-gain readout channel, saturation effects, and leakage of many shower images outside of the camera. Moreover, its scientific impact is limited because statistical uncertainties almost always dominate at ∼10 TeV due to the small photon fluxes emitted at these energies from astrophysical sources.

thumbnail Fig. 10.

Average distributions of MSCW values, in different energy bins, for the data (blue points) and simulation (red curve) in the entire post-upgrade period. They were both generated after throughput calibration. Data events were obtained from a sample of Crab Nebula observations. Due to the small total number of events, the last two energy bins are combined into a single bin with energies log10Erec(TeV) ∈ [0.3, 1.0].

5.2. Inter-telescope calibration

Each telescope that detects a γ-ray-induced shower provides an independent estimation of the primary particle energy. This enables us to test how well the throughput of each telescope is calibrated against that of the other telescopes.

There are different ways of performing such a test (Hofmann 2003; Lebohec & Holder 2003; Mitchell et al. 2016). One possibility involves selecting, for each pair of telescopes, events with a similar distance to each telescope R (in our case, within ±20%). Such events should have approximately the same reconstructed energy if the optical and camera calibration of the telescopes is appropriate. Alternatively, one can compare the energy reconstructed by a given telescope, E, against the average energy reconstructed by the array, Emean. Figure 11 illustrates the first approach (telescope pairs compared against each other) for 2019–2020.

thumbnail Fig. 11.

Inter-telescope calibration: reconstructed energies for γ-ray-like events from Crab Nebula observations, with a similar (difference less than 20%) distance from each telescope to the reconstructed shower core. The density of events in the parameter space is color-coded, with yellow values representing more events and blue values representing fewer events. A 1:1 relationship for the event energies is plotted in black, corresponding to perfect inter-telescope calibration. For compactness, we only show the results for 2019–2020.

We repeated these measurements for all IRF periods. For each period, we calculated the mean log10(Ei)−log10(Emean) and its standard deviation, shown in Fig. 12. The reconstructed energy of real γ-ray like events deviate, on average for each telescope and season, no more than ∼10% with respect to the mean energy reconstructed by the entire array. If the throughput calibration had not been implemented successfully, we would see larger differences in the energy measured by the individual telescopes. Telescopes with faster throughput degradation would provide smaller reconstructed energy estimates.

thumbnail Fig. 12.

Mean of the log10(Erec[Ti]/Emean) values for selected γ-ray-like events and its standard deviation, for each telescope and season, where Erec[Ti] is the reconstructed energy of telescope i and Emean the average reconstructed energy from the entire array, using the same weights for all telescopes. For a given epoch, the energies reconstructed by the different telescopes do not deviate on average more than ∼10% with respect to the mean energy reconstructed with the entire array.

5.3. Reconstruction of the Crab Nebula flux

The reconstructed flux and spectral shape for any measurement depends on the flux calibration of the instrument. The Crab Nebula is the closest object to a standard reference object for VHE observations as it is one of the brightest sources known that has a stable emission. Therefore, we used observations of this source for the last tests to validate both the throughput measurements and the actual implementation of the throughput calibration. This study was performed using the same Crab Nebula dataset and the same cuts as used in Sect. 4.3.4. Two figures summarizing the results of this study are obtained:

– Light curves with daily binned fluxes above 200 GeV for BDT-based moderate cuts (see Fig. 13). Datasets corresponding to the different IRF sets are plotted with different colors. Flux points versus time are shown in the left panel and a histogram of the flux points is presented in the right panel. As can be seen from the figure, every season shows an average flux compatible within ±1σ with the flux of the Crab Nebula reported in Meagher (2015). In addition, the spread of the flux points (standard deviation σ over the mean μ of the distribution) shown in the right panels remains on the level of ≲10% for the selected cuts. This value is significantly lower than the one obtained if no throughput calibration is applied. In that case, the spread would be ≳15%. The relationship between the energy threshold of the light curve and flux dispersion is shown in Fig. 14.

thumbnail Fig. 13.

Reconstructed flux of the Crab Nebula above 200 GeV, using 1-day bins and IRFs that correctly match the instrument throughput for each period. Results are obtained applying moderate cuts to runs with mean elevation > 50°. Left panel: each color represents an IRF period. The blue dot-dashed curve represents the reference Crab Nebula spectrum of Meagher (2015) integrated above 200 GeV, horizontal solid black lines represent the season-average fluxes, while dashed horizontal curves show the standard deviation of the fluxes for each season. Right panel: shown in gray is the distribution of fluxes for all seasons combined, with a fit to a Gaussian shape as a solid black line. The dashed black curve shows the equivalent result when throughput changes are not taken into account in the IRFs. The vertical blue dashed line shows the reference Crab Nebula flux from Meagher (2015).

thumbnail Fig. 14.

Flux dispersion (σ/μ) as a function of threshold energy for light curves similar to the one of Fig. 13, using moderate cuts. Taking into account the throughput evolution brings down the statistical uncertainties from ∼15% baseline to ∼10%.

– Energy spectra for the different periods, using the same cuts (Fig. 15), showing simultaneously for each period the energy spectrum obtained using the uncorrected IRFs (blue open circles), the energy spectral points for the corrected IRFs (filled red circles) and, for comparison purposes, the published spectrum from the Fermi-LAT 3FHL catalog Ajello et al. (2017) and the Crab Nebula spectrum published by the VERITAS Collaboration in Meagher (2015) (orange curved line), which is based mostly on data from seasons that were minimally affected by the throughput degradation.

thumbnail Fig. 15.

Collection of spectral energy distributions obtained with moderate cuts for each IRF period. Purple open diamonds represent measurements of the Crab Nebula spectrum from the 3FHL (Ajello et al. 2017), the orange curve shows the Crab Nebula spectrum published by VERITAS in Meagher (2015), open blue points show the fluxes that would be obtained if throughput is not calibrated during the reconstruction of γ-ray showers and instead the MC model from 2012 is employed. Finally, red solid points show the spectral points that are obtained once the pulse signals are calibrated to generate correct IRFs for each period.

6. Systematic uncertainties

The estimation of systematic uncertainties affecting the reconstructed γ-ray energies and fluxes covered in this section is based on calibration measurements, MC simulations, and reasonable assumptions. Overall, we could identify the following components of the systematic uncertainty budget impacting the throughput calibration and the determination of precise IRFs:

– Accuracy and stability (time and temperature dependence) of the photo-electron response, including the precision of the pixel-to-pixel relative gain correction. Absolute pixel gains are corrected through high-voltage adjustments (“flat-fielding”) every year at the beginning of the VERITAS observing season, in September or October. This contributes to reduce the photo-electron response drift of the PMTs over long periods of time. Relative gains likely have a small contribution to the total systematic uncertainty if the appropriate Flasher runs are used in the analysis.

– QE of the PMTs and collection efficiency at the first dynode. After several years of operations, a comparison of used and unused PMTs shows no significant evolution of these parameters over time.

– Mirror reflectivity and its wavelength dependence after correction for degradation.

– The optical PSF, determined by mirror alignment accuracy and checked once a month. Mirror facets are re-aligned after detection of significant deviation from the expected position.

Independently of the throughput calibration, the overall systematic uncertainty budget of the VERITAS instrument is affected by:

– The effect of broken pixels and electronic channels, which are not modeled in the MC simulations. The fraction of unavailable pixels per camera is typically less than 5%.

– The Winston cone efficiency, not covered in our study but included in the simulations, with an estimated contribution to the optical throughput uncertainty of ∼4%.

– The treatment of shadowing elements on the telescopes in the MC simulations, including the camera housing and support structure. Smaller elements like cross bars are not included in the simulation model. The impact of these simplifications introduces a systematic uncertainty of less than 1%.

– The systematic limit of the VERITAS pointing monitor, which is less than 20 arc seconds, much smaller than the optical PSF of the telescopes. Its contribution to the systematic uncertainty on the photon flux is therefore < 3%.

– The ability to calibrate the linearity of the readout and the transition from high to low gain using flasher pulses at eight (fifteen starting from 2016) different levels of brightness. The pulse shape of the low-gain channel is roughly twice as broad than that for high gain and exhibits a complex broadening for extremely bright signals from nearby showers with energies of tens of TeV. This complex behavior is not accurately reproduced in the current MC simulations, leading to an increase of the systematic uncertainties for the highest energies. The throughput changes discussed in this paper lead additionally to a change of pulse brightness required to activate the low-gain chain. Our correction method, which cannot account for this hardware effect, leads to a mismatch in the activation point between the simulations and the real data.

– The analysis and reconstruction pipelines, Eventdisplay and VEGAS. While the code development is independent, the analysis and reconstruction techniques employed in both pipelines are similar and they use the same MC simulations to produce the IRFs. Based on a long-term comparison of scientific results produced with both tools, we estimate a 10% systematic uncertainty due to the analysis pipeline used.

In general, the total systematic uncertainties are dominated by uncertainties on telescope throughput, covered by this work, and the approximations in the modeling of the atmosphere above the observatory.

As for the latter, the development of extensive air showers is determined by the vertical density, temperature, and humidity profiles. The intensity of Cherenkov light is influenced by the amount of absorption and scattering on atmospheric molecules or aerosols. While VERITAS does not operate during the warmest summer months, and two seasonal atmospheric profiles are used to correct for the large atmospheric changes, the systematic error due to inaccurate atmospheric models is approximately 10 − 15%. This includes day-to-day variability of the atmosphere, which is monitored by a combination of a weather station, a commercial ceilometer, and three infrared pyrometers. To limit its impact on the total systematic errors, observing periods of inferior quality (e.g., due to clouds) are flagged and removed from the analysis for most publications.

The total systematic uncertainty is estimated by summing quadratically the individual components listed in Table 2, ignoring possible correlations. As a caveat, some of the mentioned sources of systematic uncertainty are energy dependent and introduce an error on the reconstructed spectral slope, resulting in larger uncertainties for the steepest spectra. The total systematic uncertainty on the absolute flux level is estimated to be ∼25% for a γ-ray source with a spectral index of ≈2.5. Starting from ∼10 TeV, where the calibration of the low- and high-gain channels and saturation effects begin to be important, an additional 5 − 10% would have to be added in quadrature. Similarly, we estimate the systematic uncertainty on the spectral index to be ±0.2 for sources with Crab Nebula-like spectra. For sources with steeper spectra, the corresponding photon flux is dominated by the emission at the lowest energies, where many of the systematic uncertainty components that we mentioned become most relevant. In addition, the impact of the energy scale errors on the absolute flux become significantly larger and their estimation requires a case-by-case study, beyond the scope of this document.

Table 2.

Summary of systematic uncertainty estimations.

Additional sources of systematic uncertainties not directly linked to the telescope performance have not been included in this work. Nonetheless, some might have relevant implications in long-lived ground-based astronomical installations, including VERITAS. A good example of this is the long-term evolution of NSB (Massey & Foltz 2000) due to an increased human activity over time and changes in street lamp technology: from sodium lamps to LED-based illumination, likely brighter at short wavelengths Sánchez de Miguel et al. (2017), where IACTs are most sensitive. Such long-term variations are likely more evident in the direction of largely populated areas, for example, Tucson (to the north) and Nogales (to the southwest) areas, and might have a significant impact during observations at low elevations. Since VERITAS simulations are produced for different NSB levels and IRFs are interpolated to match the NSB level of each data run, this effect is corrected to first order in the standard analysis.

7. Conclusions

After almost 15 years of operations, VERITAS performance has changed due to a combination of hardware upgrades (relocation of T1, camera and camera electronics upgrades), aging of the different components, and maintenance duties, such as recoating of the most degraded mirror facets. With this work, we aim to document the calibration efforts that have been carried out by the VERITAS Collaboration during this time.

As a first step, we described in Sect. 3 the different approaches that are used to monitor the behavior of the instrument, measuring the gains of each pixel, the relative evolution of the average camera gain over time, and the changes of reflectivity and optical throughput due to telescope aging. The described methodology, now well defined, will continue to be used for the upcoming years of VERITAS operations and we are confident it will serve for other experiments as well.

In Sect. 4, we detailed the implementation of the proposed throughput calibration method in our software pipelines. It is based on correction factors for the optical throughput or reflectivity ri and the average gain of the camera gi, which we combined into a total throughput factor ti. We showed how the application of the throughput factors to the measured signals of the simulated events could be used to produce throughput-calibrated IRFs for VERITAS. These response functions can be used to analyze real showers and derive the corrected particle shower energy and source fluxes. Finally, using the time-dependent response functions obtained, we evaluated the impact of the throughput changes on the performance of VERITAS using different metrics such as the sensitivity and the energy threshold.

In order to validate this calibration method, we performed extensive tests, detailed in Sect. 5, to ensure that all VERITAS telescopes are properly calibrated against each other and that the geometrical parameters of simulated and real shower events are comparable. In addition, we checked the stability of the flux reconstruction using a multiyear sample of Crab Nebula observations, often assumed to be a reference object in the very-high-energy range.

Finally, in Sect. 6 we discussed the main systematic uncertainties on the energy scale, fluxes, and spectral indices that affect the analysis and reconstruction of data from the VERITAS telescopes.

As a final note, this work is based on direct measurements of the gains and reflectivity of the dish as an alternative to the study of local muons or the measurement of the cosmic electron spectrum. Both have been proposed and are currently being actively discussed (Parsons et al. 2016; Gaug et al. 2019) as possible calibrators for the upcoming Cherenkov Telescope Array (CTA). CTA will be comprised of arrays several times bigger than that of VERITAS, and have much tighter requirements on systematic uncertainties. It is clear that having independent methods to calibrate the response of the instrument to incident light can only help to reduce their individual limitations and systematic uncertainties.


1

A notable exception are flaring events such as GRBs, for which the rate of γ-rays may exceed the rate of background events for the first few seconds after analysis cuts have been applied (Acciari et al. 2019).

Acknowledgments

This research is supported by grants from the US Department of Energy Office of Science, the US National Science Foundation and the Smithsonian Institution, by NSERC in Canada, and by the Helmholtz Association in Germany. This research used resources provided by the Open Science Grid, which is supported by the National Science Foundation and the US Department of Energy’s Office of Science, and resources of the National Energy Research Scientific Computing Center (NERSC), a US Department of Energy Office of Science User Facility operated under Contract No. DE-AC02-05CH11231. We acknowledge the excellent work of the technical support staff at the Fred Lawrence Whipple Observatory and at the collaborating institutions in the construction and operation of the instrument. M.N. acknowledges the Young Investigators Program of the Helmholtz Association for support during the period of the project.

References

  1. Acciari, V. A., Beilicke, M., Blaylock, G., et al. 2008, ApJ, 679, 1427 [NASA ADS] [CrossRef] [Google Scholar]
  2. Acciari, V. A., Ansoldi, S., Antonelli, L. A., et al. 2019, Nature, 575, 455 [CrossRef] [Google Scholar]
  3. Aharonian, F., Akhperjanian, A. G., Aye, K. M., et al. 2004, Astropart. Phys., 22, 109 [NASA ADS] [CrossRef] [Google Scholar]
  4. Ajello, M., Atwood, W. B., Baldini, L., et al. 2017, ApJS, 232, 18 [Google Scholar]
  5. Archambault, S., Hanna, D., & Griffin, S. 2013, 33rd International Cosmic Ray Conference, 0379 [Google Scholar]
  6. Archambault, S., Chernitsky, G., Griffin, S., & Hanna, D. 2021, Astropart. Phys., 128, 102556 [NASA ADS] [CrossRef] [Google Scholar]
  7. Berge, D., Funk, S., & Hinton, J. 2007, A&A, 466, 1219 [CrossRef] [EDP Sciences] [Google Scholar]
  8. Bessell, M. S. 1990, PASP, 102, 1181 [NASA ADS] [CrossRef] [Google Scholar]
  9. Chalme-Calvet, R., de Naurois, M., & Tavernet, J. P. 2014, Proceedings of the Atmospheric Monitoring for High-Energy Astroparticle Detectors (AtmoHEAD) Conference, Saclay (France), June 10–12, 2013 [Google Scholar]
  10. Cogan, P. 2008, Int. Cosm. Ray Conf., 3, 1385 [NASA ADS] [Google Scholar]
  11. Davies, J. M., & Cotton, E. S. 1957, Sol. Energy, 1, 16 [NASA ADS] [CrossRef] [Google Scholar]
  12. de Naurois, M., & Mazin, D. 2015, C. R. Phys., 16, 610 [NASA ADS] [CrossRef] [Google Scholar]
  13. Fleury, P. 1991, Int. Cosm. Ray Conf., 2, 595 [NASA ADS] [Google Scholar]
  14. Garoli, D., Rodriguez De Marcos, L. V., Larruquert, J. I., et al. 2020, Appl. Sci., 10, 7538 [Google Scholar]
  15. Gaug, M. 2006, PhD Thesis, Autonomous University of Barcelona, Spain [Google Scholar]
  16. Gaug, M., Fegan, S., Mitchell, A. M. W., et al. 2019, ApJS, 243, 11 [NASA ADS] [CrossRef] [Google Scholar]
  17. Gazda, E., Nguyen, T., Otte, N., & Richards, G. 2016, J. Instrum., 11, P11015 [NASA ADS] [CrossRef] [Google Scholar]
  18. Hanna, D. 2008, Int. Cosm. Ray Conf., 3, 1417 [Google Scholar]
  19. Hanna, D., McCann, A., McCutcheon, M., & Nikkinen, L. 2010, Nucl. Instrum. Meth. Phys. Res. A, 612, 278 [CrossRef] [Google Scholar]
  20. Heck, D., Knapp, J., Capdevielle, J. N., Schatz, G., & Thouw, T. 1998, CORSIKA: A Monte Carlo Code to Simulate Extensive Air Showers (Germany: Forschungszentrum Karlsruhe GmbH) [Google Scholar]
  21. Hillas, A. M. 1985, Int. Cosm. Ray Conf., 3, 445 [NASA ADS] [Google Scholar]
  22. Hillas, A. M., & Patterson, J. R. 1990, J. Phys. G: Nucl. Part. Phys., 16, 1271 [NASA ADS] [CrossRef] [Google Scholar]
  23. Hofmann, W. 2003, Astropart. Phys., 20, 1 [NASA ADS] [CrossRef] [Google Scholar]
  24. Hofmann, W., Jung, I., Konopelko, A., et al. 1999, Astropart. Phys., 12, 135 [NASA ADS] [CrossRef] [Google Scholar]
  25. Holder, J. 2005, Int. Cosm. Ray Conf., 5, 379 [NASA ADS] [Google Scholar]
  26. Holder, J., Atkins, R. W., Badran, H. M., et al. 2006, Astropart. Phys., 25, 391 [NASA ADS] [CrossRef] [Google Scholar]
  27. Kieda, D. 2011, Int. Cosm. Ray Conf., 9, 14 [Google Scholar]
  28. Kieda, D. B. 2013, Int. Cosm. Ray Conf., 33, 1124 [Google Scholar]
  29. Krause, M., Pueschel, E., & Maier, G. 2017, Astropart. Phys., 89, 1 [NASA ADS] [CrossRef] [Google Scholar]
  30. Krawczynski, H., Carter-Lewis, D. A., Duke, C., et al. 2006, Astropart. Phys., 25, 380 [NASA ADS] [CrossRef] [Google Scholar]
  31. Krennrich, F., Blaylock, G., Bradbury, S. M., et al. 2007, J. Phys. Conf. Ser., 60, 34 [NASA ADS] [CrossRef] [Google Scholar]
  32. Lebohec, S., & Holder, J. 2003, Astropart. Phys., 19, 221 [NASA ADS] [CrossRef] [Google Scholar]
  33. Li, T. P., & Ma, Y. Q. 1983, ApJ, 272, 317 [CrossRef] [Google Scholar]
  34. López Paredes, B., Araújo, H. M., Froborg, F., et al. 2018, Astropart. Phys., 102, 56 [CrossRef] [Google Scholar]
  35. MacLeod, A. 2007, Master’s Thesis, McGill University Libraries, Canada [Google Scholar]
  36. Maier, G., & Holder, J. 2017, Int. Cosm. Ray Conf., 301, 747 [CrossRef] [Google Scholar]
  37. Massey, P., & Foltz, C. B. 2000, PASP, 112, 566 [NASA ADS] [CrossRef] [Google Scholar]
  38. McCann, A., Hanna, D., Kildea, J., & McCutcheon, M. 2010, Astropart. Phys., 32, 325 [NASA ADS] [CrossRef] [Google Scholar]
  39. Meagher, K. 2015, Int. Cosm. Ray Conf., 34, 792 [Google Scholar]
  40. Meyer, M. 2005, High Energy Gamma-Ray Astronomy, eds. F. A. Aharonian, H. J. Völk, & D. Horns, AIP Conf. Ser., 745, 774 [NASA ADS] [CrossRef] [Google Scholar]
  41. Mirzoyan, R., Garczarczyk, M., Hose, J., & Paneque, D. 2007, Astropart. Phys., 27, 509 [NASA ADS] [CrossRef] [Google Scholar]
  42. Mitchell, A. M. W., Parsons, R. D., Hofmann, W., & Bernlöhr, K. 2016, Astropart. Phys., 75, 1 [NASA ADS] [CrossRef] [Google Scholar]
  43. Otte, A. N., Gebremedhin, L., Kaplan, K., & Long, D. 2011, Int. Cosm. Ray Conf., 9, 247 [Google Scholar]
  44. Parsons, R. D., Hinton, J. A., & Schoorlemmer, H. 2016, Astropart. Phys., 84, 23 [NASA ADS] [CrossRef] [Google Scholar]
  45. Perkins, J. S., Maier, G., & The VERITAS Collaboration 2009, ArXiv e-prints [arXiv:0912.3841] [Google Scholar]
  46. Roache, E., Irvin, R., Perkins, J., et al. 2007, 30th International Cosmic Ray Conference (ICRC2017), 1397 [Google Scholar]
  47. Roache, E., Irvin, R., Perkins, J. S., et al. 2008, Int. Cosm. Ray Conf., 3, 1397 [NASA ADS] [Google Scholar]
  48. Rovero, A. C., Buckley, J. H., Fleury, P., et al. 1996, Astropart. Phys., 5, 27 [NASA ADS] [CrossRef] [Google Scholar]
  49. Sánchez de Miguel, A., Aubé, M., Zamorano, J., et al. 2017, MNRAS, 467, 2966 [CrossRef] [Google Scholar]
  50. Speckmayer, P., Höcker, A., Stelzer, J., & Voss, H. 2010, J. Phys.: Conf. Ser., 219, 032057 [NASA ADS] [CrossRef] [Google Scholar]
  51. Tyler, J. 2013, Int. Cosm. Ray Conf., 33, 3096 [Google Scholar]
  52. Vacanti, G., Fleury, P., Jiang, Y., et al. 1994, Astropart. Phys., 2, 1 [NASA ADS] [CrossRef] [Google Scholar]
  53. VERITAS Collaboration (Nagai, T., et al.) 2007, ArXiv e-prints [arXiv:0709.4517] [Google Scholar]
  54. Weekes, T. C. 1996, Space Sci. Rev., 75, 1 [NASA ADS] [CrossRef] [Google Scholar]
  55. Winston, R. 1974, Sol. Energy, 16, 89 [NASA ADS] [CrossRef] [Google Scholar]
  56. Winston, R. 1976, Light Collectors in Cylindrical Geometry, United States, https://www.osti.gov/biblio/7129508 [Google Scholar]
  57. Zitzer, B. 2013, Int. Cosm. Ray Conf., 33, 3076 [NASA ADS] [Google Scholar]

Appendix A: Reference mirror reflectivities used in the simulations

thumbnail Fig. A.1.

Mean telescope reflectivity used to produce the baseline MC simulations in the current instrument configuration of VERITAS, and the reference IRFs discussed throughout this work.

The reflectivities of the mirror facets used in the simulation of the VERITAS telescopes are based on measurements made in 2011, after T1 was relocated to its current position in the post-upgrade array layout. Together with the average camera gains of Section 3.1, they are part of the current detector model of VERITAS, defined in April 2012 and used as a reference throughout this document. The reflectivities were measured from 260 nm to 700 nm for several mirrors facets that were located on the top, middle, and bottom of the dish of each telescope. The measured values were then averaged to get a mean reflectivity representative of the entire dish.

Appendix B: Detailed MC-Data comparison

During the validation of the proposed method to calibrate the throughput, we performed extensive tests to check the agreement between data and the modified MC simulations. One of the most sensitive tests in VERITAS to detect possible discrepancies between the measured Cherenkov signals and the corresponding simulated showers is to check the agreement in the MSCW and MSCL parameters. Both are shown, for each IRF period, in Figures B.1 and B.2.

thumbnail Fig. B.1.

Comparison between the MSCW values of simulated γ-ray showers (red curves) and the ones for real γ-like events (blue dots). Each row corresponds to one IRF period and each column to a energy bin (in log-scale, with the energy measured in TeV).

thumbnail Fig. B.2.

Comparison between the MSCL values of simulated γ-ray showers (red curves) and the ones for real γ-like events (blue dots). Each row corresponds to one IRF period and each column to a energy bin (in log-scale, with the energy measured in TeV).

Appendix C: Stability of spectral parameters

As discussed in Section 2.3, the Crab Nebula is often used as a source to benchmark the performance of γ-ray instruments. It is not only one of the brightest sources in the sky, but also visible from both hemispheres. The Crab Nebula has a hard spectrum which reaches energies of a few tens of TeV with photon fluxes that are detectable by VERITAS. Finally, it is thought to be stable in year-timescales both in spectral shape and emitted luminosity, at least in the very-high-energy band. Using eight one-year observation campaigns of the Crab Nebula for the entire post-upgrade period of VERITAS, we could evaluate the performance of the proposed throughput calibration method.

Figure C.1 shows as a time series the run-wise values of the normalization flux and the spectral index of the source, evaluated at 1 TeV. It was estimated by analysing each Crab Nebula run that survived quality cuts and had more than 4 σ excess. Then, the resulting run-wise spectra were fitted locally with a power-law over the small energy range of 0.4 − 6 TeV. A histogram representation of the values of Figure C.1 can be seen in Figure C.2.

thumbnail Fig. C.1.

Run-wise measurement of the normalization flux at 1 TeV and the spectral index of the Crab Nebula between 0.6 and 4 TeV, including all IRF periods starting from the upgrade of VERITAS. Only runs with a detection significance of at least 4 σ are included.

thumbnail Fig. C.2.

Histogram of run-wise measurements of normalization and spectral index of Crab Nebula between 0.6 and 4 TeV including all IRF periods starting from the upgrade of VERITAS. Only runs with a detection significance of at least 4 σ are included.

All Tables

Table 1.

IRF periods for the current VERITAS instrument configuration.

Table 2.

Summary of systematic uncertainty estimations.

All Figures

thumbnail Fig. 1.

Extensive air shower reconstructed by the VERITAS telescopes. The shower images for each camera have been integrated over time and cleaned using a two-level filter. Dead and disabled channels are shown in dark gray and brown respectively. The signal registered by each PMT is color-coded, with red colors representing higher signal than blue tones. An approximate geometrical reconstruction of the shower core location is illustrated with red dashed lines and a red star. The upper right plot shows the trace of one of PMT (#255) of T1 for reference. There, the signal registered for each sample of 2 ns is plotted in red and the integration window (six samples) is shown in shadowed blue.

In the text
thumbnail Fig. 2.

Time-dependent changes of photo-statistic gains normalized to the nominal absolute gains used in the baseline MC model for the current detector model. This detector model was constructed in 2012, and reproduces the characteristics of VERITAS after the PMT camera upgrade. Black points show the average gain factors for each instrument epoch. Gray points show the individual unfiltered gain factor values per telescope and daily flasher run. Blue, orange, green and red points show the result of a median filter of the individual gain factors. Curves of the same colors represent a spline interpolation of the filtered values. Even though spline interpolation can locally diverge when there is no data available, we note that this occurs only during Summer breaks, where no data is taken in VERITAS, therefore it is not a concern for this study. Error bars represent statistical uncertainties.

In the text
thumbnail Fig. 3.

Time-dependent changes of the reflectivity factors obtained from WDR measurements. Black points show the average reflectivity factors for each instrument epoch. Blue, orange, green, and red points show the individual reflectivity factor measurements for each telescope. Curves of the same colors represent a spline interpolation, which removes outliers. The first point (early 2012) is extracted from the simulations and it serves as the reference for the calculation of the reflectivity factors, therefore adopting a value of 1 as reflectivity factor. We note that a value of the reflectivity factor slightly greater than 1 just means that the reflectivity at that time was slightly higher than at the time where simulations were produced.

In the text
thumbnail Fig. 4.

Time-dependent change of the throughput factors calculated from photo-statistic gain factors and WDR reflectivity factors. Error bars show only statistical uncertainties.

In the text
thumbnail Fig. 5.

Ratio between image parameters derived from MC simulations for the season 2019–2020 vs season 2012–2013a (Telescope T1 only, winter simulations, noise level of 100 MHz, and no size cuts). The vertical axis of each panel shows the number of events within a given ratio bin.

In the text
thumbnail Fig. 6.

Relative changes in energy reconstruction induced by the evolution of the throughput of the instrument. The ratio of throughput-calibrated energies with respect to the uncorrected energies is presented in two formats: As 2D lookup tables, with the ratio as function of impact distance and size of the simulated events (bottom panels); curves with the distance-averaged lookup-tables, showing the ratio as a function only of size (top panels).

In the text
thumbnail Fig. 7.

Evolution of the effective area for different instrument epochs, for a zenith angle of observation of 20°, noise level of 100 MHz, moderate cuts and a minimum of two images per event.

In the text
thumbnail Fig. 8.

Energy threshold of the analysis estimated from simulated events and using moderate cuts, zenith angle of 20°, and noise level of 100 MHz. It is defined as the energy for which the effective area becomes 10% of the maximum effective area.

In the text
thumbnail Fig. 9.

Differential sensitivity with 50 h of VERITAS observations calculated using different low zenith angle Crab Nebula data, with winter atmospheric profile conditions. The same cuts as in Fig. 7 are used. The sensitivity is defined in terms of the flux of the Crab Nebula (in Crab units, CU) assuming it is a reference astrophysical object at VHE energies.

In the text
thumbnail Fig. 10.

Average distributions of MSCW values, in different energy bins, for the data (blue points) and simulation (red curve) in the entire post-upgrade period. They were both generated after throughput calibration. Data events were obtained from a sample of Crab Nebula observations. Due to the small total number of events, the last two energy bins are combined into a single bin with energies log10Erec(TeV) ∈ [0.3, 1.0].

In the text
thumbnail Fig. 11.

Inter-telescope calibration: reconstructed energies for γ-ray-like events from Crab Nebula observations, with a similar (difference less than 20%) distance from each telescope to the reconstructed shower core. The density of events in the parameter space is color-coded, with yellow values representing more events and blue values representing fewer events. A 1:1 relationship for the event energies is plotted in black, corresponding to perfect inter-telescope calibration. For compactness, we only show the results for 2019–2020.

In the text
thumbnail Fig. 12.

Mean of the log10(Erec[Ti]/Emean) values for selected γ-ray-like events and its standard deviation, for each telescope and season, where Erec[Ti] is the reconstructed energy of telescope i and Emean the average reconstructed energy from the entire array, using the same weights for all telescopes. For a given epoch, the energies reconstructed by the different telescopes do not deviate on average more than ∼10% with respect to the mean energy reconstructed with the entire array.

In the text
thumbnail Fig. 13.

Reconstructed flux of the Crab Nebula above 200 GeV, using 1-day bins and IRFs that correctly match the instrument throughput for each period. Results are obtained applying moderate cuts to runs with mean elevation > 50°. Left panel: each color represents an IRF period. The blue dot-dashed curve represents the reference Crab Nebula spectrum of Meagher (2015) integrated above 200 GeV, horizontal solid black lines represent the season-average fluxes, while dashed horizontal curves show the standard deviation of the fluxes for each season. Right panel: shown in gray is the distribution of fluxes for all seasons combined, with a fit to a Gaussian shape as a solid black line. The dashed black curve shows the equivalent result when throughput changes are not taken into account in the IRFs. The vertical blue dashed line shows the reference Crab Nebula flux from Meagher (2015).

In the text
thumbnail Fig. 14.

Flux dispersion (σ/μ) as a function of threshold energy for light curves similar to the one of Fig. 13, using moderate cuts. Taking into account the throughput evolution brings down the statistical uncertainties from ∼15% baseline to ∼10%.

In the text
thumbnail Fig. 15.

Collection of spectral energy distributions obtained with moderate cuts for each IRF period. Purple open diamonds represent measurements of the Crab Nebula spectrum from the 3FHL (Ajello et al. 2017), the orange curve shows the Crab Nebula spectrum published by VERITAS in Meagher (2015), open blue points show the fluxes that would be obtained if throughput is not calibrated during the reconstruction of γ-ray showers and instead the MC model from 2012 is employed. Finally, red solid points show the spectral points that are obtained once the pulse signals are calibrated to generate correct IRFs for each period.

In the text
thumbnail Fig. A.1.

Mean telescope reflectivity used to produce the baseline MC simulations in the current instrument configuration of VERITAS, and the reference IRFs discussed throughout this work.

In the text
thumbnail Fig. B.1.

Comparison between the MSCW values of simulated γ-ray showers (red curves) and the ones for real γ-like events (blue dots). Each row corresponds to one IRF period and each column to a energy bin (in log-scale, with the energy measured in TeV).

In the text
thumbnail Fig. B.2.

Comparison between the MSCL values of simulated γ-ray showers (red curves) and the ones for real γ-like events (blue dots). Each row corresponds to one IRF period and each column to a energy bin (in log-scale, with the energy measured in TeV).

In the text
thumbnail Fig. C.1.

Run-wise measurement of the normalization flux at 1 TeV and the spectral index of the Crab Nebula between 0.6 and 4 TeV, including all IRF periods starting from the upgrade of VERITAS. Only runs with a detection significance of at least 4 σ are included.

In the text
thumbnail Fig. C.2.

Histogram of run-wise measurements of normalization and spectral index of Crab Nebula between 0.6 and 4 TeV including all IRF periods starting from the upgrade of VERITAS. Only runs with a detection significance of at least 4 σ are included.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.