Calibrating photometric redshifts with intensity mapping observations
Imaging surveys of galaxies will have a high number density and angular resolution yet a poor redshift precision. Intensity maps of neutral hydrogen (HI) will have accurate redshift resolution yet will not resolve individual sources. Using this complementarity, we show how the clustering redshifts approach, proposed for spectroscopic surveys can also be used in combination with intensity mapping observations to calibrate the redshift distribution of galaxies in an imaging survey and, as a result, reduce uncertainties in photometric redshift measurements. We show how the intensity mapping surveys to be carried out with the MeerKAT, HIRAX and SKA instruments can improve photometric redshift uncertainties to well below the requirements of DES and LSST. The effectiveness of this method as a function of instrumental parameters, foreground subtraction and other potential systematic errors is discussed in detail.
Photometric redshift surveys are an economic way of building up a detailed map of the large scale structure of the Universe. By imaging large swathes of the sky, it is possible to construct catalogues of individually resolved galaxies with high number density (and therefore a low “shot” noise). The trade-off for such a large number of objects is the inability to obtain accurate redshift measurements for individual objects. Thus, photometric redshift surveys are orders of magnitude less resolved in the radial direction than the sparser spectroscopic redshift surveys. The uncertainty in the individual redshifts and in the overall galaxy redshift distribution can severely degrade the constraining power of such datasets for cosmology.
Galaxies cluster to form the cosmic web, and one expects structures in the galaxy distribution to be spatially correlated with structures in any other tracer of the dark matter density. For example, if one has an imaging survey of galaxies (where redshifts are poorly resolved) and a spectroscopic catalog (where redshifts are well resolved), they should have non-trivial cross-correlations; in particular, structures in the imaging survey should be mirrored in the spectroscopic survey. A natural step is to use these cross-correlations so that the precise redshift measurements of the spectroscopic survey can be used to sharpen the photometric redshifts in the imaging survey, or at least calibrate its redshift distribution. These types of methods have been advocated in Newman (2008); Benjamin et al. (2010); Matthews and Newman (2010); Schmidt et al. (2013); Ménard et al. (2013); van Daalen and White (2017), and employed in the analysis of several datasets (e.g. Rahman et al. (2016a); Choi et al. (2016); Scottez et al. (2016); Rahman et al. (2016b); Johnson et al. (2017)).
One does not necessarily have to use a catalogue of resolved sources to follow this rationale. In particular, if one can accurately map out, in redshift, any tracer of the dark matter, it can in principle be used to improve redshift measurements in a sister imaging survey. A notable example is that of an unresolved map of neutral hydrogen, HI, through a technique known as intensity mapping (Battye et al., 2004; McQuinn et al., 2006; Chang et al., 2008; Wyithe and Loeb, 2008; Loeb and Wyithe, 2008; Peterson et al., 2009; Bagla et al., 2010; Battye et al., 2013; Masui et al., 2013; Switzer et al., 2013; Bull et al., 2015). Radio observations at a GHz or below will map out the distribution of neutral hydrogen out to redshifts, or higher. The neutral hydrogen traces the large scale structure of the dark matter and thus, inevitably, will be correlated with any other tracer. Maps of HI will be exquisitely resolved in the frequency domain and therefore will map out the density distribution, in detail, in redshift. Although intensity mapping observations will not resolve individual objects, they will be able to achieve sufficient angular resolution for cosmological studies (although this statement depends on the observing mode).
In this paper we explore the use of HI intensity mapping to calibrate photometric redshift surveys. In particular we show that forthcoming intensity mapping experiments such as those undertaken by MeerKAT Fonseca et al. (2017), HIRAX Newburgh et al. (2016) and the SKA Santos et al. (2015) can be used to reduce the uncertainties related to photo- systematics well below the requirements currently posited by the DES and LSST surveys, thus improving final constraints on cosmological parameters. We structure this paper as follows: in Section II we describe the formalism and discuss, in detail, various aspects of the instrumental and observational models we are assuming. Section III presents our results as a function of experimental configuration and foreground uncertainties. In Section IV we discuss the prospects of using such a method and compare with other proposals currently being developed. The appendices present a number of calculations which are essential for the models considered here.
ii.1 Clustering-based photo- calibration
Consider two galaxy samples with redshift distributions (), and let be the harmonic coefficients of their projected overdensity of counts on the sky. Their cross-correlation is given by:
where is the matter power spectrum, is the radial comoving distance, is a spherical Bessel function, is the cross-noise power spectrum between samples and , is the linear bias of the -th sample and we have neglected redshift-space distortions and all other sub-dominant contributions to the observed power spectrum. In the Limber approximation - where - this expression simplifies to:
For the purposes of this discussion, the most important feature of Equation 3 is the fact that the amplitude of the cross-correlation is proportional to the overlap between the redshift distributions of those samples. This is especially relevant if one of the samples has good radial resolution, in which case it can be split into narrow bins of redshift. The cross-correlations of all narrow bins with the other sample will therefore trace the amplitude of its redshift distribution, and can effectively be used to constrain it. This is illustrated in Fig. 1, which shows the cross-power spectrum between a Gaussian photo- bin of width and a set of narrow redshift bins ().
Note also that Eq. 3 implies that the redshift distribution and the redshift-dependent galaxy bias of the photometric sample are completely degenerate in this method, and therefore additional information is needed in order to separate both quantities (e.g. including prior information or lensing data). Since this is an inherent problem of the method, and not specific to the case of intensity mapping, we will simply assume that is a sufficiently smoothly-varying function of , thus treating IM and spectroscopic surveys on an equal footing. The more complicated biasing scheme that arises on small scales also prevents the use of those modes to constrain Schmidt et al. (2013), and therefore one must be conservative when deciding the range of scales to include in the analysis.
Different recipes have been formulated to carry out this kind of analysis, such as the optimal quadratic estimator method of McQuinn and White (2013). The forecasts presented here will interpret the redshift distribution (in a parametric or non-parametric form) as a set of extra nuisance parameters, on which we will carry out the Fisher matrix analysis described in Section II.5. Thus, even though our results will be optimistic in as much as the Fisher matrix saturates the Rao-Cramer bound, they will account for all correlations between redshift distribution parameters and with the cosmological parameters, as well as the presence of redshift-space distortions and magnification bias (effects that have been overseen in previous works).
For the purposes of estimating the ability of future surveys to calibrate photometric redshift distributions through cross-correlations, we will always consider an individual redshift bin for a photometric sample with unknown distribution, together with a set of overlapping narrow redshift bins of spectroscopic galaxies or intensity mapping observations. Let be the overall true redshift distribution of the photometric sample, and let be the conditional distribution for a photo- given the true redshift . Then, the redshift distribution in a photo- redshift bin with bounds is given by
In what follows we will consider two degrees of complexity in terms of describing the unknown redshift distribution:
We will assume Gaussian photo-s with a given variance () and bias :
and we will assume that the uncertainty in the redshift distribution is fully described by and .
We will use a non-parametric form for , given as a piecewise function with a free amplitude for each spectroscopic redshift bin.
Our assumed fiducial value for and , as well as the binning scheme used are described in Section II.2.
We finish this section by noting that the use of cross-correlations with spectroscopic surveys or intensity mapping observations for photo- calibration is not limited to the measurement of the redshift distribution of a given galaxy sample, but that they can also be used to improve the precision of photometric redshift estimates for individual galaxies (e.g. Jasche and Wandelt (2012)). Although we leave the discussion of this possibility for future work, we describe a Bayesian formalism for this task in Appendix A.
ii.2 Photometric redshift surveys
This section describes the model used here for a LSST-like photometric redshift survey. As in Alonso and Ferreira (2015), we base our description of the number density of sources and their magnification bias on the measurements of the luminosity function of Gabasch et al. (2006), with -corrections computed with kcorrect Blanton and Roweis (2007). We assume a magnitude cut of in the band, corresponding to the so-called “gold” sample LSST Collaboration et al. (2009). Unlike Alonso and Ferreira (2015), and for simplicity, we will consider a single galaxy population, instead of splitting it into “red” and “blue” sources. The resulting redshift distribution is shown by the solid black line in Figure 2.
The photometric redshift requirement for the gold sample as stated in the LSST science book are , with a goal of . Here we have taken a conservative estimate, assuming a standard deviation . We then split the full sample into redshift bins with a width given by , where is the photo- variance at the bin centre. This binning scheme is chosen to reduce the correlation between bins induced by the tails of the photo- distribution, and results in the 15 redshift bins shown in Fig. 2 (where the redshift distributions are computed with Eq. 4). Our fiducial photo- model will assume biased Gaussian distributions, fully determined by and .
ii.3 Intensity mapping
|25K||25 K||50 K|
|10000 h||4000 h||h|
|15 m||13.5 m||6 m|
|range||350-1050 MHz||600-1050 MHz||400-800 MHz|
Intensity mapping (IM) is a novel observational technique that circumvents the long integration times needed to obtain reliable spectroscopic redshifts for individual objects through an approach that is transverse to that used by photometric surveys. The idea (McQuinn et al., 2006; Wyithe and Loeb, 2008; Peterson et al., 2009; Switzer et al., 2013) is to observe the unresolved combined emission of many line-emitting sources in a relatively wide pixel at different frequencies. The signal-to-noise ratio of the corresponding line emission is much stronger than that of the individual sources, and thus, combining the intensity measured across the sky and relating the intensity observed at a given frequency to the rest-frame wavelength of the emission line it is possible to produce three-dimensional maps of the density of the line-emitting species. This technique is particularly appealing for isolated spectral lines, as is the case of the 21cm line caused by the spin-flip transition in neutral hydrogen atoms (HI), and thus HI intensity mapping has been proposed as an ideal method to cover vast volumes at relatively low cost.
A number of experiments have been proposed to carry out IM measurements of the baryon acoustic oscillation scale, such as BINGO Battye et al. (2012), CHIME Newburgh et al. (2014), FAST Bigot-Sazy et al. (2016), HIRAX Newburgh et al. (2016), SKA Santos et al. (2015) and Tianlai Chen (2011). The different instrumental approaches to IM can be broadly classified into two camps:
Interferometers: the sky emission is measured by a set of antennas, and the measurements of pairs of antennas separated by a given baseline are cross-correlated to produce the measurement of an angular Fourier mode with scale (where is the observed wavelength). The intensity map is then reconstructed by combining pairs with different baselines.
Single-dish: in this case the sky emission is measured and auto-correlated by individual antennas. A band-limited intensity map with a resolution is then produced by varying the antenna pointing, where is the antenna diameter.
The expressions for the noise power spectrum for both cases are derived in Appendix C, and can be summarized as:
Here is the system temperature, given as a combination of instrumental and sky
temperature (see Appendix C), is the sky fraction covered
by the observations, is the antenna efficiency
where is the total number of independent baselines.
Given their expected full overlap with LSST, we will consider here the two main currently envisaged southern-hemisphere intensity mapping experiments: SKA (and its pathfinder, MeerKAT) and HIRAX.
MeerKAT and the SKA
MeerKAT is the 64-dish precursor to the mid-frequency component of the SKA. MeerKAT is comprised of 13.5 metre dishes and will operate between GHz using three separate receivers. Although, it will predominantly used as an interferometer, and as such only be sensitive to relatively small spatial scales, there is a proposed project to use MeerKAT in single-dish mode Fonseca et al. (2017). If such a mode of operation is viable, then MeerKAT will become an extremely efficient intensity-mapping facility operating at frequencies that allows the detection of Hi to . Indeed, a proposed open-time survey would provide a several thousand square degree sky survey over the Dark Energy Survey and/or Kilo-degree Survey areas, which will provide excellent visible wavelength coverage.
In the 2020s, MeerKAT will be enhanced by the addition of 130, 15 metre dishes to form the mid-frequency SKA. Operating at similar frequencies to MeerKAT, the additional 130 dishes will provide much more sensitivity for all science aims, and is capable of carrying out a deg intensity mapping survey Santos et al. (2015).
As such, both MeerKAT and the SKA will provide a unique view on the Hi Universe, and as we will show, can enhance the cosmological science with the LSST with cross-correlations.
The Hydrogen Intensity mapping and Real-time Analysis eXperiment (HIRAX) is a proposed close-packed radio array comprising 1024 six metre dishes disposed in a grid and operating at 400-800 MHz. The telescope will be located on the South African Karoo site, which has very low levels of RFI in this band, and provides an ideal location to overlap in sky coverage with other planned southern sky cosmological surveys. The large collecting area and field-of-view provide excellent sensitivity and mapping speed, with the high density of short baselines allowing for sensitive measurements of the baryon acoustic oscillation (BAO) scale in the cosmic HI distribution from redshift 0.8 to 2.5, which in turn will provide competitive constraints on dark energy (Newburgh et al., 2016). HIRAX will make high signal-to-noise maps of 21cm intensity fluctuations over 15,000 sq degrees (taken to overlap fully with LSST) on cosmological scales of interest, with the relatively high frequency resolution (1024 channels over the 400 MHz bandwidth) allowing for accurate redshift calibration of 21cm intensity maps. This makes it ideal for calibration of LSST photometric redshifts through the cross-correlation technique.
Generic IM experiment
Besides SKA and HIRAX we will also explore the capabilities of a generic intensity mapping experiment in terms of photo- calibration. The performance of a given experiment is roughly determined by three quantities:
The range of angular scales over which the noise power spectrum is low enough to probe the cosmological HI emission. This range can be characterized by the minimum and maximum baselines and .
The noise level (normalized by the bandwidth ) on this range of scales. For a fixed integration time, this is determined by the system temperature and the observed sky area .
Here we will model the effects of the minimum and maximum baselines as a sharp and an inverse-Gaussian cutoff respectively. Thus, our model for the angular noise power spectrum is:
where and is 1 if and 0 otherwise. Note that by definition has units of . For comparison, the equivalent values of these parameters that roughly reproduce the noise curves for HIRAX are:
One of the main obstacles that HI intensity mapping must overcome to become a useful cosmological tool is the presence of galactic and extragalactic foregrounds several orders of magnitude larger than the 21cm cosmological signal Santos et al. (2005); Wolz et al. (2014). Under the assumption that foregrounds are coherent in frequency (as opposed to the cosmic signal tracing the density inhomogeneities along the line of sight), these foreground sources can be in principle efficiently removed using component-separation methods Wolz et al. (2014); Shaw et al. (2015); Alonso et al. (2015). However, instrumental imperfections, such as frequency-dependent beams or polarisation leakage, can generate foreground residuals with a non-trivial frequency structure that could strongly bias cosmological constraints from 21cm data alone. In any case, the removal of frequency-smooth components will introduce large uncertainties on the large-scale radial modes of the 21cm fluctuations.
Here we have introduced the effect of foregrounds by including an extra component, , in the sky model for HI accounting for foreground residuals. Thus we will assume that the measured harmonic coefficients at a given frequency are given by:
where and are the true cosmological signal and the instrumental noise contribution. We will model as an almost-correlated component with a cross-frequency power spectrum given by
Here and parametrise the amplitude of the foreground residuals and their
distribution on different angular scales, and describes their mean frequency
dependence. Finally, parametrises the characteristic frequency scale over which
foregrounds are decorrelated. When including the effects of foregrounds (Section
III.3) we will also marginalize over .
For and we will use the fiducial values and ,
corresponding to galactic synchrotron emission Santos et al. (2005); Planck Collaboration et al. (2016), and we will set , large enough for the
residuals to dominate the equal- power spectrum
Effectively, this extra component cancels the constraining power of all radial modes of the 21cm fluctuations with comoving radial wavenumbers below a scale , related to through
The effect of foregrounds on the ability to constrain redshift distributions can be readily understood as loss of information in the space. In the flat-sky approximation, and on linear scales, the angular power spectrum between two tracers and of the matter density can be computed as:
where we have again ignored the effect of redshift-space distortions and:
Here is the linear growth factor and and are the linear bias and selection function for the -th tracer. Let us assume that is a photometric redshift bin and is a narrow intensity mapping frequency shell with comoving width centered at . Assuming and the bias to be slowly-varying functions of we obtain:
Now, assuming that has support over a wide range of redshifts, corresponding to a comoving width , its Fourier transform () will only have support over wavenumbers . Since the Bessel function provides support over all values of , and under the assumption that , the total number of modes that contribute to is bound by . Since foreground contamination will mostly affect large radial modes, eventually a large fraction of this -range becomes dominated by foreground uncertainties and stops contributing efficiently to the overall signal-to-noise ratio, thus degrading the final constraints on any model parameter.
We finish this Section by noting that, as described in Section II.1, the main constraining power for photo- calibration comes from the cross-correlation of the photometric and spectroscopic samples. Since the photometric sample would not suffer from foreground contamination, these cross-correlations are very robust against foreground biasing, which makes photo- calibration an ideal application of IM.
ii.4 Spectroscopic surveys
In order to showcase the possibility of calibrating redshift distributions through cross-correlation with future intensity mapping experiments we will compare their forecast performance against that of the most relevant future spectroscopic suveys:
The Dark Energy Spectroscopic Instrument (DESI) Levi et al. (2013) is a spectroscopic galaxy survey planned to cover from its northern-hemisphere location at Kitt Peak National Observatory. We assume an area overlap of with LSST, and we model the number density and clustering bias of the two galaxy samples considered here (Luminous Red Galaxies and Emission Line Galaxies) as done in Font-Ribera et al. (2014).
The Wide Field Infrared Survey Telescope (WFIRST) Spergel et al. (2013) is a future space observatory in the infrared that will measure redshifts for objects over . The deep nature of WFIRST will make it ideal to calibrate the LSST redshift distribution at high redshifts. We model the number density and bias of the WFIRST sample as in Font-Ribera et al. (2014), and we assume a full overlap with LSST ().
Figure 3 shows the redshift distributions for galaxies detected by these three experiments.
ii.5 Forecasting formalism
Our formalism will distinguish between two types of tracers of the density field:
Spectroscopic: tracers whose redshift distribution is well known. This would correspond to tracers with good radial resolution such as a narrow redshift bin of spectroscopic sources or an intensity map in a narrow frequency band, as well as other tracers with a well-known window function, such as a CMB lensing map.
Photometric: tracers whose redshift distribution is unknown or uncertain. This would correspond to e.g. a photometric-redshift bin, a radio continuum survey or a map of the Cosmic Infrared Background.
Let us start by considering a set of sky maps corresponding to a number of tracers, and let be the corresponding vector of maps expressed in a given basis. In the following sections we will assume that is stored in terms of spherical harmonic coefficients and that it takes the form , where is a photometric tracer and is a set of spectroscopic tracers. For the moment, however, we will keep the discussion general.
Assuming that is Gaussianly distributed with zero mean and covariance , its log-likelihood is given by:
Now let be a set of parameters modelling , including (but not limited to) the parameters describing the photometric redshift distribution. A maximum-likelihood estimator for can be defined by using an iterative Newton-Raphson method to minimize Eq. 15. This is described in Tegmark et al. (1998); Bond et al. (1998); McQuinn and White (2013), and yields the iterative algorithm:
where, in Eq. 16 there is an implicit summation over , the sub-index implies differentiation with respect to , is the Fisher matrix, is the th iteration of the solution for and the previous iteration is used to compute and in the second term. Note that we have simplified a pure Newton-Raphson iteration by taking the ensemble average of the likelihood Hessian (i.e. the Fisher matrix). Furthermore, in the case where the likelihood is well-approximated by a Gaussian, is the covariance matrix of the . Eq. 16 is the basis of the method proposed in McQuinn and White (2013) (with a number of simplifications) and used in Johnson et al. (2017) to constrain the redshift distribution of galaxies in the KiDS survey.
In our case, we mainly care about the uncertainty in the redshift distribution parameters included in the , and therefore we will simply estimate the Fisher matrix . In the case where is a set of spherical harmonic coefficients with power spectrum , is given by
where we have approximated the effects of a partial sky coverage by scaling the number of independent modes per by the sky fraction . The form of the power spectra for the different tracers considered in this work is given in Appendix B.
As explicitly shown in Eq. 17, smaller-scale modes carry a higher statistical weight (proportional to ), and would in principle dominate the redshift distribution constraints. The smallest scales are, however, dominated by theoretical uncertainties from non-linearities in the evolution of the density field and the galaxy-halo connection, and therefore a multipole cutoff must be used to contain the constraining power of systematics-dominated modes. In this paper we use a redshift-dependent cutoff defined as follows. Let be the mean redshift of a given redshift bin, and let be the variance of the linear density field at that redshift on modes with wavenumber :
We then define the cutoff scale as , where satisfies for some choice of . In what follows we will use a fiducial threshold , corresponding to , and we will study the dependence of our results on this choice. Besides this choice of , we will also impose a hard cutoff for all galaxy-survey and intensity-mapping tracers of (thus, in reality, ).
In order to forecast for the ability of future experiments to constrain photometric redshift distributions, in the following sections we will use the formalism described in Section II.5 with a data vector given by , where is a photometric redshift bin and are a set of overlapping redshift bins for a spectroscopic tracer (either an intensity mapping experiment or a spectroscopic galaxy survey). The number , width and redshift range of the spectroscopic redshift bins is chosen in order to adequately sample the changes in the photometric redshift distribution. We choose the redshift bin width to be of the photo- uncertainty , which governs the variability of the redshift distribution (i.e. each redshift interval of is sampled in 3 points). In order to sample the tails of the distribution we then define the redshift range of the set of spectroscopic bins as , where and are the edges of the photometric redshift bin. The number of spectroscopic redshift bins is then defined in terms of these numbers.
The model parameters in the following sections will be given by:
All of the parameters needed to determine the redshift distribution (, or the amplitude in different spectroscopic bins, depending on the case).
Two overall clustering bias parameters, and , corresponding to the bias of the photometric and spectroscopic tracers.
We will also include two cosmological parameters: the fractional matter density and the amplitude of scalar perturbations , in in order to account for the possible cosmology dependence of the results.
We will change this setup in Section III.5, where we will explore the impact of the achieved constraints on the photo- parameters on the final cosmological constraints. In this section will correspond to the 15 photometric redshift bins for LSST, for both galaxy clustering and weak lensing (i.e. 30 sets of spherical harmonics). Likewise will contain the cosmological parameters as well as all the baseline photo- parameters ( and for all redshift bins), with priors corresponding to the constraints found in the preceding sections.
iii.1 Baseline forecasts
Using the formalism described above, and in the simplified scenario of Gaussian photo-s, we present, in the left panel of Figure 4, the forecast constraints on the photo- bias () and variance () for the key intensity mapping experiments introduced in Section II.3. In this and all subsequent plots, the thin black solid line shows the LSST requirement of Newman et al. (2015); de Putter et al. (2014), while the thin dashed line corresponds to the nominal requirement for the Dark Energy Survey () The Dark Energy Survey Collaboration (2005). The coloured bars in these and all subsequent plots show the redshift ranges corresponding to the proposed frequency bands of the three IM experiments explored here.
Two key features must be noted in this figure: first, the uncertainties grow steeply at low redshifts. This is due to the reduced number of modes available in that regime, associated with the smaller comoving volume and the impact of non-linearities on lower values of . The latter effect is especially severe for HIRAX, given its inability to measure angular modes larger than its beam size. Note, however, that this regime lies outside the proposed frequency ranges for both HIRAX () and SKA ().
Secondly, the ratio between and stays roughly constant (). This is compatible with the expected ratio between the uncertainties associated to the maximum-likelihood estimates of the mean and standard deviation of a Gaussian distribution from a finite number of samples (). This result holds for most of the cases explored here (see Section III.3 for an exception), and thus we have omitted the curves for in most of the subsequent figures.
The right panel of Fig. 4 compares the constraints achievable by IM
experiments with those forecast for the spectroscopic surveys described in Section
II.4. We see that both SKA and HIRAX would be able to satisfy the LSST
requirements over the redshift range of interest. The SKA precursor MeerKAT would fall short
except at low redshifts. However, the shorter-term timeline of MeerKAT (2018 onward)
would make it an ideal experiment to prove the viability of this technique in cross-correlation
with the Dark Energy Survey (DES) The Dark Energy Survey
Collaboration (2005), particularly in the light of the
proposed intensity mapping surveys Fonseca et al. (2017) targeting a full overlap with
As discussed in Bull et al. (2015), the dish size of SKA is not ideal for cosmological observations, since it is not large/small enough to resolve the angular BAO scale sufficiently well in either single-dish or interferometric modes, although single-dish observations are able to address important science cases such as primordial non-Gaussianity Camera et al. (2013); Alonso and Ferreira (2015). Small scales carry a larger statistical weight, however, and it is not clear that a single-dish strategy would also be ideal for the purposes of photo- calibration. This is explored in Figure 5, which shows the constraints on achievable with single-dish (dashed lines) and interferometric (dash-dotted lines) observations for SKA (red) and MeerKAT (yellow). The constraints from a joint auto- and cross-correlation analysis are shown as solid lines, and correspond to the results reported here. We see that, in the case of SKA, the single-dish mode outperforms the interferometer up to , when a sufficiently large number of usable modes enter the regime probed by the latter. This suggests that, if simultaneous single-dish and interferometric observations proved to be unfeasible, the photo- calibration requirements could still be met by using either mode in different redshift ranges.
The performance of this method at low redshift depends crucially on the prescription used to isolate the effect of non-linearities. Here we have done this in terms of the threshold rms variance defined in Eq. 18 for a fiducial value of , corresponding to at . Figure 6 shows the result of relaxing or tightening this criterion. The effect on SKA and MeerKAT is only moderate, since these experiments gather most of their sensitivity from the large, linear scales in auto-correlation mode. HIRAX, on the other hand, loses sensitivity more rapidly as the scale of non-linearities removes a larger fraction of the available modes. Nevertheless, even for (corresponding to at ) the LSST calibration requirements are satisfied in the redshift range corresponding to the HIRAX frequency band.
iii.2 Dependence on experimental parameters
We have so far quantified the potential of currently-proposed experimental configurations for photo- calibration. The aim of this section is to identify the optimal instrumental specifications for this task.
We start by exploring the balance between noise level and sky fraction, varying the total overlapping area of the three experiments explored in the previous section (SKA, MeerKAT and HIRAX) keeping the total observation time fixed. The result is presented in Figure 7, which shows the achievable constraints on when reducing the sky area by successive factors of 2. We find that, although the results are almost insensitive to the reduction in , larger sky areas are always preferred, a reflection of the fact that the constraints are dominated by cosmic variance rather than noise.
As we have discussed in the previous section, the key drawback of single-dish experiments is their inability to access angular scales smaller than the beam size (with their higher statistical weight). Conversely, interferometers are unable to cover scales larger than that probed by their smallest baseline, and therefore they have access to a limited number of reliable (i.e. mildly non-linear) modes. Using the generic instrument parametrization given by Eq. 8, and the fiducial parameters corresponding to HIRAX, Figure 8 explores these issues.
The upper panel shows the constraints achievable by a single-dish experiment for different dish sizes. Dishes of at least 15m (corresponding to the SKA case) are needed in order to achieve the LSST requirements at all redshifts, while a dish (e.g. such as the Green Bank Telescope Masui et al. (2013)) would be able to achieve constraints similar to those of HIRAX. The largest currently planned experiment is FAST Bigot-Sazy et al. (2016), with a dish size of 500m.
The lower panel of the figure shows the performance of an interferometer as a function of the minimum baseline . In this case the main effect is the loss, at lower redshifts, of the large, mildly non-linear scales. A maximum baseline of at most 12m is needed in order to calibrate redshift distributions below with sufficient accuracy.
As we discussed in Section II.3.4, the main effect of radio foregrounds for 21cm observations is to make large radial scales (i.e. modes with smaller than some ) useless. We have parametrised this by introducing an extra component corresponding to foreground residuals characterised by an amplitude and a frequency correlation length (see Section II.3.4 and Eq. 10 for a full description). We set to a value large enough to dominate the emission on radial scales larger than the comoving length corresponding to (see Eq. 11), and study the final constraints as a function of .
Figure 9 shows the result of this analysis: while sufficiently coherent foregrounds () do not qualitatively modify the final results for the three experiments under consideration, correlation lengths smaller than would result in a fast degradation of the performance for photo- calibration. In particular, the associated loss of -space volume would prevent MeerKAT and SKA from reaching the calibration requirements for DES and LSST. The peformance of HIRAX would also be severely compromised by foreground contamination, although it would still be able to reach the required constraints within its proposed frequency range.
These results pose the question of how uncorrelated we can expect foreground residual to be. For reference, raw foreground components are constrained to have correlation lengths of Santos et al. (2005). On the other hand, more complicated residuals arising from leaked polarised synchrotron would exhibit a much richer frequency structure caused by Faraday rotation, with correlation lengths at high frequencies () decreasing for longer wavelengths Jelić et al. (2010); Shaw et al. (2015); Alonso et al. (2015). An exquisite instrumental calibration will therefore be necessary in order to optimise the scientific output of 21cm experiments. It is worth noting that the comoving scale corresponding to ( at ) is similar to the cut suggested by Shaw et al. (2015), although the expontial form assumed here for the power spectrum of the foreground residuals (Eq. 10) extends the degrading effect into larger wavenumbers.
A further complication comes in the form of the so-called “foreground wedge”, produced by the long time-delay contribution of foregrounds from antennas with far side-lobe responses Datta et al. (2010); Vedantham et al. (2012); Liu et al. (2014); Seo and Hirata (2016). As proven in Parsons et al. (2012); Pober et al. (2014), this effect makes the region of -space defined by:
liable to foreground contamination. This is the so-called “horizon” wedge, and corresponds to the case where foreground contamination can be caused by the sources in the horizon picked up by very far sidelobes. Under optimistic assumptions, however, we can consider the case where this effect extends only up to foreground sources located in the outer fringes of the primary beam, in which case the size of the wedge is reduced to the so-called primary-beam wedge Pober et al. (2014), given by . The lower panel of Figure 9 shows that the LSST photo- calibration requirements are still met after accounting for the loss of -space to the primary beam wedge.
iii.4 Generalized redshift distributions
Even though the simple parametrization of photo- systematic uncertainties in terms of and allows us to easily compare the performance of different experiments in terms of photo- calibration, in a realistic scenario we would like to calibrate generic redshift distributions without assuming a particular parametrization.
This is typically done by promoting the amplitude of the redshift distribution of the photometric sample in each narrow redshift interval to a free parameter that can be constrained from the cross-correlation with the spectroscopic survey. In this section we explore this scenario for the same redshift bins considered in the previous sections.
For this we use the method proposed by McQuinn and White (2013). In essence, this method is equivalent to the formalism outlined in Section II.5, where the free parameters considered are the amplitudes of the photometric redshift distribution. The method is further simplified in McQuinn and White (2013) to make it applicable to the analysis of real data using the following approximations:
All power spectra are computed using the Limber approximation. This implies (among other things) that all cross-correlations between non-overlapping redshift bins are neglected.
The contributions from RSDs and magnification bias are not included in the model for the angular power spectra.
No marginalization over cosmological or other nuisance parameters is carried out.
The auto-correlation of photometric sources does not contain information about their redshift distribution.
We have adopted these same assumptions here to simplify the discussion.
Figure 10 shows the constraints achievable by different IM experiments
and spectroscopic surveys on the generalized form of the redshift distribution for three
photometric redshift bins centered around and 2.6
Note that the uncertainties on the amplitude of the redshift distribution found this way can be translated into uncertainties on the two parameters and used to characterize this distribution in the previous section by performing a simple 2-parameter likelihood analysis for the model in Eqs. 4 and 5. We find that, using this procedure, we recover constraints on and that are a factor worse than in the optimal case. This can be understood in terms of the simplifying assumptions adopted here, such as neglecting the information encoded in cross-bin correlations and in the auto-correlation of the photometric sample. In all cases, however, we recover the same relative performance between different experiments in terms of .
iii.5 Impact on cosmological constraints
In order to study the impact of photo- calibration on the final cosmological constraints we have run a Fisher matrix analysis using the formalism described in Section II.5 with the specifications for LSST described in Section II.2. In this case we consider a set of 54 free parameters:
3 nuisance parameters for each of the 15 redshift bins: the balaxy bias , the photo- bias and the photo- scatter .
9 cosmological parameters: the fractional density of cold dark matter , the fractional density in baryons , the normalized local expansion rate , the amplitude and tilt of primordial scalar perturbations and , the optical depth to reionization , the equation of state of dark energy in the - parametrization and the sum of neutrino masses .
In order to pin down the early-universe parameters, we also include information from a hypothetical ground-based Stage-4 CMB experiment Abazajian et al. (2016) using the specifications assumed in Calabrese et al. (2016) and complemented by Planck at low multipoles.
For the photo- nuisance parameters we then add Gaussian priors corresponding to the uncertainties on and found using the procedure described in the previous sections for the different experiments considered in this paper. The results of this exercise are displayed in Figure 11, which shows the constraints on the most relevant late-universe parameters: the dark energy equation of state parameters, and , and the sum of neutrino masses. The results shown correspond to the contours in the absence of photo- uncertainties (dashed black line), with photo- systematics constrained through cross-correlation with a 21cm experiment, in this case HIRAX (red ellipse) and in the absence of external data for photo- calibration (blue ellipse).
It is important to stress that the overall constraints on these parameters forecast here depend heavily on the survey specifications assumed (e.g. photo- model and uncertainties), as well as the scales included in the analysis, a point where we have tried to be very conservative. Thus, the results shown in Figure 11 should not be taken to represent the final constraints achievable by LSST. The main result shown in this figure is the relative improvement on the final constraints after photo- calibration, which should be more robust to these considerations.
Photo- calibration improves the constraints on each of these parametes by a factor , and the dark energy figure of merit by a factor . Furthermore, we find that the level of calibration achievable through cross-correlations with intensity mapping experiments is nearly equivalent to the case of perfect calibration. Equivalent results are found for SKA and HIRAX, as well as for the combination of DESI, Euclid and WFIRST, as could be expected from the results displayed in the right panel of Figure 4.
In this paper we have shown that intensity maps of the HI emission can be used to improve the scientific output of photometric redshift surveys. By exploiting the cross-correlations between imaging surveys of galaxies and HI maps, we find that it is possible to optimally calibrate the photometric redshift distributions. This is made clear when assessing improvements in constraints on cosmological parameters: in Figure 11 we see that the FoM using this method is effectively as good as having perfect calibration. This also means that, with the aid of future IM experiments it should be possible to, for example, improve the LSST equation of state figure of merit by approximately a factor of 5.
This approach is promising, but it is important to highlight some of the limitations that have to be overcome if IM is to be used successfully in this context. For a start, it will be important to be able to deal with foreground contamination. In essence, as we have discussed, one can model the effect of foregrounds as a source of noise that cancels the information contained in long-wavelength radial modes. This effect is therefore greatly dependent on how coherent foreground residuals are in frequency. We can expect foregrounds to be highly correlated in intensity, with correlation lengths of order . However, instrumental effects such as polarization leakage or frequency-dependent beams could spoil this coherence and lead to significant losses in the coverage of the - plane. We estimate that residuals with correlation lengths (as expected for polarised synchrotron emission, and corresponding to scales at ) would significantly degrade the ability of 21cm maps to calibrate redshift distributions efficiently. We have furthermore shown that the calibration requirements of future photometric surveys can also be matched after accounting for the so-called “foregrounds wedge”. Finally, it is worth noting that one of the main strength of the method is its reliance on cross-correlations between the spectroscopic and photometric samples, and that this cross-correlation should be very robust against systematic biases caused by foreground contamination.
If we are to use IM to calibrate future redshift surveys, then we need to make sure that the observational set up satisfies minimum requirements and to do so, we have explored its dependence on experimental parameters. We have found that, for the noise levels of currently proposed experiments, we are primarily limited by cosmic variance and therefore there is no advantage in gaining depth at the cost of sky area: it is preferable to maximize the overlap between the HI maps and the galaxy imaging survey. Furthermore, if we are to accurately capture the longer wave-length modes, we need to resort to single dish observations, a promising but as yet untested mode of observation for MeerKAT and SKA. In the long term, “” noise contributions will have to be controlled, and our analysis shows that the dishes must have a minimum size of metres. If, on the other hand, we are to use interferometric observations (a method which is more tried and tested) then we need to ensure a minimum baseline of metres to capture the large-scale angular modes. We have shown how MeerKAT, HIRAX and SKA fall well within these experimental parameter constraints.
We must also note that our analysis has been conservative in terms of the ranges of scales that add up to the constraints on photo- parameters, only including angular scales in the regime where non-linear structure formation is believed to be well understood. It may, however, be possible to use even smaller scales for the purposes of photo- calibration Ménard et al. (2013), in which case some of the conclusions drawn from this analysis could vary. In particular, the relative performance of HIRAX and SKA in terms of photo- calibration could be significantly different, owing to the higher sensitivity of the SKA interferometer to small angular scales (see Figure 13).
It is interesting to consider alternative approaches to sharpening photometric redshift measurements. Gravitational lensing of the CMB has recently been advocated as a promising approach, given its perfectly well-determined radial kernel (e.g. see Baxter et al. (2016)). We can explore this possibility by considering correlations between LSST and a CMB lensing convergence map as in Section II.5. We consider two different generations of CMB experiments: an ongoing “Stage-3” experiment, characterised by an rms noise level of , and a future “Stage-4” experiment with . In both cases we assume very optimistic configurations, with a beam FWHM of 1 arcmin and using all angular scales . We also fix all cosmological and bias parameters, and only consider varying and for each LSST redshift bin. Figure 12 shows the constraint on the photo- bias in the absence of CMB lensing data (green solid line), and using the cross-correlation with as measured by S3 (purple) and S4 (cyan) experiments. We clearly see that, even under these overly optimistic assumptions, adding CMB lensing information does not significantly help calibrate photo- uncertainties.
Experiments to undertake intensity mapping of HI are an ongoing effort, and will give us an extremely promising new window on the Universe. We have argued that they will not only contribute in their own right to the further understanding of the large-scale structure of the Universe, but will also help improve the scientific returns from a plethora of up and coming optical surveys.
The authors would like to thank Elisabeth Krause, Mario Santos and Anže Slosar for useful comments and discussions. DA is supported by the Science and Technology Facilities Council and the Leverhume and Beecroft Trusts. PGF acnowledges support from Science Technology Facilities Council, the European Research Council and the Beecroft Trust. MJJ acknowledges support from The SA SKA project and the Oxford centre for astrophysical surveys which is funded through generous support from the Hintze Family Charitable Foundation. KM acknowledges support from the National Research Foundation of South Africa (Grant Number 98957).
Appendix A Individual clustering redshifts
The idea of using clustering information to constrain the redshifts of individual objects of a given sample has been considered before in the literature, and shown to yield interesting results even in the absence of spectroscopic data Jasche and Wandelt (2012). Here we outline the steps that should be taken to include intensity mapping information in this formalism.
Our aim is to find the most general expression for the posterior distribution of the true redshifts of a set of galaxies for which we only have photometric data and an overlapping intensity mapping survey. We start by considering a data vector consisting of:
: galaxy positions.
: galaxy magnitudes.
: a map of the perturbation in the HI density across angles and redshift
For each galaxy we want to estimate a redshift , so let be a vector containing all those redshifts. We want to study the posterior probability . Let us start by noting that, in the standard models of large-scale structure, both and the galaxy distribution can be thought of as being biased and noisy representations of the underlying matter overdensity . Sampling the galaxy redshifts could then also give us information about , and therefore it’s worth considering the joint distribution .
One can study this distribution by iteratively sampling the two conditional distributions:
We outline these two steps below.
Conditional density distribution. We start by noting that, if the true redshifts are known, then the photometric redshifts given by the magnitudes are of no use in constraining the overdensity field, and therefore
where is the galaxy overdensity uniquely defined by the galaxy angular coordinates and redshifts. can then be decomposed using Bayes’ theorem as:
where, following the same philosophy as above, we have considered that , since is just a noisy and biased realization of . All that remains is then to model the conditional distributions and , which is by no means a cursory matter, but something that can certainly be accomplished in the regime of validity of structure formation models.
Conditional redshift distribution. Under the assumptions that galaxies are Poisson-distributed over the (biased) matter overdensity, and that the photometric redshift errors are independent of the environmental density, it is possible to show (e.g. Jasche and Wandelt (2012)) that the galaxy redshifts can be sampled individually with the distribution:
where is the galaxy overdensity along the angular direction of each galaxy, and is the prior photo- distribution.
Appendix B Angular power spectra
This section describes the theoretical models used for the angular power spectra entering the computation of the Fisher matrix (Eq. 17). The cross-power spectrum between two tracers of the cosmic density field, and , can be estimated as:
where is the power spectrum of the primordial curvature perturbations and is the window function for tracer , containing information about the different contributions to the total anisotropy in that tracer and about its redshift distribution.
In the case of galaxy clustering and intensity mapping, and neglecting contributions from magnification bias and large-scale relativistic effects, is given by:
where and are the expansion rate and radial comoving distance at redshift respectively, is the source redshift distribution, and and are the transfer functions of the matter overdensity and velocity divergence fields. Note that, even though we include the effect of non-linearities using the non-linear transfer function for (through the prescription of Takahashi et al. (2012)), we only introduce the effect of redshift-space distortions at the linear level, and only consider a deterministic linear bias . This is, nevertheless, a more rigorous treatment than has been used in the literature, and the procedure used to mitigate the effect of non-linearities described in Section II.5 should minimize the corresponding impact on the forecasts presented here.
For galaxy shear tracers of weak lensing, the expression for the window function is:
where is the transfer function for the sum of the two metric potentials in the Newtonian gauge.
Appendix C Noise power spectrum for intensity mapping experiments
This section derives the expression for the noise power spectra of single-dish experiments and interferometers presented in Eq. 6. Similar derivations have been provided before in the literature (e.g. Bull et al. (2015)), but we include this calculation here for completeness.
Throughout this section we will use a flat-sky approach, where angles on the sky are represented by a 2D Cartesian vector . In this approximation, the spherical harmonic transform of a field becomes a simple 2D Fourier transform:
We will also simplify the derivation by writing integrals as Riemann sums. For instance, the Fourier transform above will read:
Note that, with this normalization, the definition for the power spectrum of a stochastic field is
The flux at angular position measured by a single dish is the sky intensity integrated over the instrumental beam :
where . Inserting the definition 29 in the expression above, and using the orthogonality relation , one can relate the Fourier components of and as
The power spectrum for is then related to that of as . Assume now that is purely white noise with a per-pointing rms flux , such that its power spectrum is simply flat with an ampitude:
where is the angular separation between pointings, is the per-pointing rms temperature fluctuation, is the effective collecting area of the dish, is the channel frequency bandwidth, is the integration time per pointing, is the total observed sky area and is the total integration time.
Substituting this into the expression for and relating the intensity to a brightness temperature through the Rayleigh-Jeans law (), we obtain the temperature noise power spectrum:
where is the observed sky fraction, we have considered the possibility of having independent dishes, and we have defined the quantity . Note that, for a circular aperture telescope, , where is the dish diameter, and therefore is the ratio of the effective to total dish area, which we have labelled “efficiency” in Eq. 6.
The visibility observed by a pair of antennas separated by a baseline is
Transforming this to Fourier space we find:
where, in the last step, we have used the small-angle approximation . The variance of is therefore given by
where is the number of baselines in a volume in -space.
In temperature units, the noise variance per visibility is given by . Relating baselines to Fourier coefficients as and recalling the definition of power spectrum (Eq. 30), the noise power spectrum in temperature is then given by
where is the total number of pointings. Relating to the number density of physical baselines, and defining we recover the expression for the noise power spectrum of interferometers in Eq. 6.
Comparison with spectroscopic surveys
Converting the angular maps in different frequency channels into a three-dimensional map of the HI overdensity, we can relate the 3D noise power spectrum to its angular counterpart as:
where is the comoving angular diameter distance to redshift and is the average 21cm temperature. This can then be directly compared with the shot-noise power spectrum of spectroscopic surveys, given by , where is the 3D density of sources. The left panel of Figure 13 shows the 3D noise power spectrum at as a function of the transverse wavenumber for the three IM experiments (HIRAX, SKA and MeerKAT) and the three next-generation spectroscopic surveys (DESI, Euclid and WFIRST) considered here.
- is defined as the ratio of the effective to real antenna area.
- We use a pivot scale an a privot frequency , as in Santos et al. (2005).
- Note that the photo- calibration requriements, defined in terms of the degradation of the final constraints, should be less stringent for DES
- Note that the first and last bins would lie outside the proposed frequency range of HIRAX
- J. A. Newman, Astrophys. J. 684, 88-101 (2008), arXiv:0805.1409 .
- J. Benjamin, L. van Waerbeke, B. Ménard, and M. Kilbinger, MNRAS 408, 1168 (2010), arXiv:1002.2266 [astro-ph.CO] .
- D. J. Matthews and J. A. Newman, Astrophys. J. 721, 456 (2010), arXiv:1003.0687 [astro-ph.CO] .
- S. J. Schmidt, B. Ménard, R. Scranton, C. Morrison, and C. K. McBride, MNRAS 431, 3307 (2013), arXiv:1303.0292 .
- B. Ménard, R. Scranton, S. Schmidt, C. Morrison, D. Jeong, T. Budavari, and M. Rahman, ArXiv e-prints (2013), arXiv:1303.4722 .
- M. P. van Daalen and M. White, ArXiv e-prints (2017), arXiv:1703.05326 .
- M. Rahman, A. J. Mendez, B. Ménard, R. Scranton, S. J. Schmidt, C. B. Morrison, and T. Budavári, MNRAS 460, 163 (2016a), arXiv:1512.03057 .
- A. Choi, C. Heymans, C. Blake, H. Hildebrandt, C. A. J. Duncan, T. Erben, R. Nakajima, L. Van Waerbeke, and M. Viola, MNRAS 463, 3737 (2016), arXiv:1512.03626 .
- V. Scottez, Y. Mellier, B. R. Granett, T. Moutard, M. Kilbinger, et al., MNRAS 462, 1683 (2016), arXiv:1605.05501 .
- M. Rahman, B. Ménard, and R. Scranton, MNRAS 457, 3912 (2016b), arXiv:1508.03046 .
- A. Johnson, C. Blake, A. Amon, T. Erben, K. Glazebrook, et al., MNRAS 465, 4118 (2017), arXiv:1611.07578 .
- R. A. Battye, R. D. Davies, and J. Weller, Mon. Not. Roy. Astron. Soc. 355, 1339 (2004), arXiv:astro-ph/0401340 [astro-ph] .
- M. McQuinn, O. Zahn, M. Zaldarriaga, L. Hernquist, and S. R. Furlanetto, Astrophys. J. 653, 815 (2006), astro-ph/0512263 .
- T.-C. Chang, U.-L. Pen, J. B. Peterson, and P. McDonald, Phys. Rev. Lett. 100, 091303 (2008), arXiv:0709.3672 [astro-ph] .
- J. S. B. Wyithe and A. Loeb, M.N.R.A.S. 383, 606 (2008).
- A. Loeb and J. S. B. Wyithe, Physical Review Letters 100, 161301 (2008), arXiv:0801.1677 .
- J. B. Peterson, R. Aleksan, R. Ansari, et al., in astro2010: The Astronomy and Astrophysics Decadal Survey, ArXiv Astrophysics e-prints, Vol. 2010 (2009) p. 234, arXiv:0902.3091 [astro-ph.IM] .
- J. Bagla, N. Khandai, and K. K. Datta, Mon. Not. Roy. Astron. Soc. 407, 567 (2010), arXiv:0908.3796 [astro-ph.CO] .
- R. A. Battye, I. W. A. Browne, C. Dickinson, G. Heron, B. Maffei, and A. Pourtsidou, M.N.R.A.S. 434, 1239 (2013), arXiv:1209.0343 [astro-ph.CO] .
- K. W. Masui, E. R. Switzer, N. Banavar, K. Bandura, C. Blake, et al., ApJL 763, L20 (2013), arXiv:1208.0331 [astro-ph.CO] .
- E. R. Switzer, K. W. Masui, K. Bandura, L.-M. Calin, T.-C. Chang, et al., MNRAS 434, L46 (2013), arXiv:1304.3712 [astro-ph.CO] .
- P. Bull, P. G. Ferreira, P. Patel, and M. G. Santos, Astrophys. J. 803, 21 (2015), arXiv:1405.1452 .
- J. Fonseca, R. Maartens, and M. G. Santos, MNRAS 466, 2780 (2017), arXiv:1611.01322 .
- L. B. Newburgh, K. Bandura, M. A. Bucher, T.-C. Chang, H. C. Chiang, et al., in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Proceedings of the SPIE, Vol. 9906 (2016) p. 99065X, arXiv:1607.02059 [astro-ph.IM] .
- M. Santos, P. Bull, D. Alonso, S. Camera, P. Ferreira, et al., Advancing Astrophysics with the Square Kilometre Array (AASKA14) , 19 (2015), arXiv:1501.03989 .
- M. McQuinn and M. White, MNRAS 433, 2857 (2013), arXiv:1302.0857 .
- J. Jasche and B. D. Wandelt, MNRAS 425, 1042 (2012), arXiv:1106.2757 .
- D. Alonso and P. G. Ferreira, Phys. Rev. D 92, 063525 (2015),