# 4.2. The datasets¶

We list and detail below the datasets already implemented and give instructions on how to include other data.

## Type Ia supernovae¶

Two types of analyses can be made with the JLA catalogue.
Here we are using the binned data consisting of distance modulus estimates at
31 points (defining 30 bins of redshift).
This dataset is select via the `JLAsimple`

entry in the list of data sources.
If you want to use the full dataset (which makes the analysis much slower since
it involves three more nuisance parameters and requires the program to invert a
740 by 740 matrix at every iteration for the calculation of the JLA
likelihood), insert `SNeJLA`

instead. Also, you need to download the full
data from http://supernovae.in2p3.fr/sdss_snls_jla/ReadMe.html or from my
copy on Google Drive.
Since the data files are too big, EPIC includes only the
`jla_likelihood_v4/data`

folder. In this case you will need the
`jla_likelihood_v4/covmat`

. Make sure to properly include it in you
installation directory.

Either way, the argument is the same data folder `jla_likelihood_v4`

.
Note that the binned dataset introduces one nuisance parameter `M`

,
representing an overall shift in the absolute magnitudes, and the full dataset
introduce other four nuisance parameters related to the light-curve parametrization.
See Betoule et al. (2014) [1] for more details.

Note

This version of EPIC supports version V4 (June 2014) of the dataset release. Current version is V6 (March 2015). An update to implement use of V6 is being considered for the near future.

## CMB distance priors¶

Constraining models with temperature or polarization anisotropy amplitudes is
not currently implemented.
However, you can include the CMB distance priors from Planck2015 with
`Planck2015_distances`

.
These data consist of an acoustic scale \(l_A\), a shift parameter
\(R\) and the physical density of baryons \(\Omega_{b0}h^2\).
You can choose between the data for \(\Lambda\text{CDM}\),
\(w\text{CDM}\) and \(\Lambda\text{CDM} + \Omega_k\) models [2], specifying
the folders `Planck2015_LCDM`

, `Planck2015_wCDM`

or
`Planck2015_LCDM+Omega_k`

, respectively.

## BAO data¶

Measurements of the baryon acoustic scales from the Six Degree Field Galaxy
Survey (6dF), the Main Galaxy Sample of Data Release 7 of Sloan Digital Sky
Survey (SDSS-MGS), the LOWZ and CMASS galaxy samples of the Baryon Oscillation
Spectroscopic Survey (BOSS-LOWZ and BOSS-CMASS), the WiggleZ Dark Energy Survey
and the distribution of the Lyman \(\alpha\) forest in BOSS (BOSS-Ly) are
compiled file `BAO-6dF+SDSS+BOSS+Lyalpha+WiggleZ.txt`

.
The subsamples `BAO-6dF+SDSS+BOSS+Lyalpha.txt`

and `BAO-6dF.txt`

are also
available, the first excluding the WiggleZ data and the second only with the
6dF data.
Since these files simply contain the redshift of the measurement, the value of
the characteristic ratio \(r_{\text{BAO}}(z) \equiv r_s(z_d)/d_V(z)\) between
the sound horizon \(r_s\) at decoupling time (\(z_d\)) and the
effective BAO distance \(d_V\) and the measurement error, it is really
simple to remove or add new measurements of this observable.

## \(H(z)\) data¶

These are the cosmic chronometer data.
30 measurements of the Hubble expansion rate \(H(z)\) at redshifts between
0 and 2, plus a \(2.4\%\) precision local measure of \(H_0\).
The values of redshift, \(H\) and the uncertainties are given in the file `thirtypointsHz.txt`

.

## \(f\sigma_8\) data¶

Large-scale structure data from redshift-space distortion and peculiar velocity measurements giving the growth rate times the RMS amplitude of matter perturbations \(f \sigma_8(z)\) can also be used. Notice that in this case you need to provide an analytic evaluation of the growth rate (for example, as \(f = \Omega_m^{\gamma}\)) in your model since this program does not evolve perturbations numerically at the present moment. See for example the implementation in Marcondes et al. (2016) [3].

## Including other data¶

You can include other data if you want.
In the simplest case, you will have measurement values and error bars
for a certain quantity, say \(f(z)\), at given points.
Put them on a text file in three columns separated by TAB or spaces, as in, for example, the `SDSS+BOSS.txt`

file:

```
0.15 4.47 0.17
0.32 8.47 0.17
0.57 13.77 0.13
```

Choose a label for this dataset and put it in the `.ini`

file in the data files section:

```
my_data my_data_file.txt
```

Next, you need to tell the code how to read this new type of dataset. In this simple case you add a function like:

```
def my_data(obsble):
return simplest_data(obsble, 'my_data', r'$f(z)$')
```

to the `load_data`

module.
The name of this function must be the
same as the label that goes in the `.ini`

file. This function, or its
more complicated equivalent, is the only change needed to be made in the
`load_data.py`

file.
In the `simplest_data`

function call, the `obsble`

variable carries
the dataset file information (the filename or directory), the second
argument is a string to be displayed indicating that your data is loaded.
The third argument is a raw string, possibly including
\(\LaTeX\) notation, to be displayed in the triangle plots
compounding the label that describes the datasets used in this analysis
when you plot together results from more than one analysis.

You then need to define the likelihood calculation
in the `likelihood`

module. At the end of the file, add to the
dictionary `allprobes`

an entry with the key `'my_data'`

(again the
same label) and the likelihood function, say `my_data_likelihood`

, as
its value. For a Gaussian likelihood, this function can be:

```
def my_data_likelihood(datapoints, cosmology):
return simple(datapoints, observables.my_f, cosmology)
```

Now, as you may have already figured out, the next step is to add the
theoretical calculation `my_f`

of your observable \(f\) to the
`observables.py`

file. Of course, this will vary with each case. Do
not forget to use the `model`

attribute of the `Cosmology`

class
object named `cosmology`

to set different calculations according to
each cosmological model you might want to use.

[1] | Betoule M. et al. “Improved cosmological constraints from a joint analysis of the SDSS-II and SNLS supernova samples”. Astronomy & Astrophysics 568, A22 (2014). |

[2] | Huang Q.-G., Wang K. & Wang S. “Distance priors from Planck 2015 data”. Journal of Cosmology and Astroparticle Physics 1512, 022 (2015). |

[3] | Marcondes R. J. F., Landim R. C. G., Costa A. A., Wang B. & Abdalla E. “Analytic study of the effect of dark energy-dark matter interaction on the growth of structures”. Journal of Cosmology and Astroparticle Physics 1612, 009 (2016). |