Likelihood Overview (Adapted from the FSSC's Cicerone)In order to analyze LAT data, it is necessary to construct the likelihood that is applicable to the LAT data, and then use this likelihood to find the best fit model parameters, including the description of a source's spectrum, its position, and even whether it exists. The likelihood L is the probability of obtaining the data given an input model. In this case, the input model is the distribution of gamma-ray sources on the sky, and includes their intensity and spectra. Binned vs Unbinned LikelihoodUnbinned likelihood analysis is the preferred method for spectral fitting of the LAT data. However, a binned analysis is provided for cases where the unbinned analysis cannot be used. For example, the memory required for the likelihood calculation scales with number of photons and number of sources. This memory usage becomes excessive for long observations of complex regions, necessitating the use of binned analysis.Model FittingAssuming that we know a source is present, we expect the best model to have the highest probability of resulting in the data, and we vary the spectral parameters until the likelihood is maximized. Note that χ2 is -2 times the logarithm of the likelihood in the limit of a large number of counts in each bin, and therefore where χ2 is a valid statistic, minimizing χ2 is equivalent to maximizing the likelihood. To fit a source's spectra:
Goodness-of-fit. Likelihood spectral fitting provides the best fit parameter values and their uncertainties, but is this a good fit?
Test Statistic. The Test Statistic is defined as TS=-2ln(Lmax,0/Lmax,1), where Lmax,0 is the maximum likelihood value for a model without an additional source (the 'null hypothesis') and Lmax,1 is the maximum likelihood value for a model with the additional source at a specified location. As can be seen, TS is a monotonically increasing function of Lmax,1, which is why maximizing TS on a grid is equivalent to maximizing the likelihood on a grid. In the limit of a large number of counts, Wilkes Theorem states that the TS for the null hypothesis is asymptotically distributed as χ2x. Note: Here χ2 is the distribution, not a value of the statistic, where x is the number of parameters characterizing the additional source. This means that TS is drawn from this distribution if no source is present, and an apparent source results from a fluctuation. Thus, a larger TS indicates that the null hypothesis is incorrect (i.e., a source really is present), which can be quantified. Note: As a basic rule of thumb, the square root of the TS is approximately equal to the detection significance for a given source. Select Data: What data should be used for source analysis?Choosing the Data to Analyze — Regions of Interest and Source Regions Assume that we are analyzing the spectrum of a single source. Because of the large point spread function at low energies (e.g., 68% of the counts will be within 3.5 degrees at 100 MeV), we want to use the counts within a region around the source. Nearby sources will contribute counts to this region, and we want to model them, i.e., to model a single source we are forced to model a handful of sources. Therefore, we need to include counts from an even larger region. For the greatest accuracy possible in modeling a single source, we should model the entire sky(!), but this is not usually feasible and, in reality, the influence of sources a very great distance away from the source will be greatly attenuated. Thus ,we include sources from a large 'Source Region' and counts from a smaller 'Region of Interest' (ROI). The positions and spectra of sources in the Source Region outside of the ROI were obtained previously; from a catalog, for example. These sources are included because of their contribution to the counts in the ROI. How we treat the sources in the ROI is under our control: we may wish to fix the parameters of sources other than the one we are studying at their catalog values, or we might want to perform a fit on the parameters of all of these sources. To summarize, we will use all of the sources in the Source Region, and determine the size of the Source Region appropriate for our needs from experience and experimentation. Recommended values. Default values of ROI+10 and ROI+5 degrees are recommended for sources dominated by ~100 MeV and ~1 GeV events, respectively, and all counts in the ROI are included. The appropriate ROI size is determined based on experience and experimentation, but the recommended default values are 20 and 15 degrees, respectively, for sources dominated by ~100 MeV and ~1 GeV events. Precomputation of Likelihood QuantitiesThe computation of the likelihood usually occurs many times. Fits are done with various model parameters fixed or with different sources present or absent. Certain quantities need be calculated only once, speeding up the repeated computation of the likelihood. Livetime CubeThe LAT instrument response functions are a function of the inclination angle, the angle between the direction to a source and the LAT normal. The number of counts that a source should produce should therefore depend on the amount of time that the source spent at a given inclination angle during an observation. This livetime quantity, the time that the LAT observed a given position on the sky at a given inclination angle, depends only on the history of the LAT's orientation during the observation and not on the source model. The array of these livetimes at all points on the sky is called the 'livetime cube.' As a practical matter, the livetime cubes are provided on a healpix grid on the sky and in inclination angle bins (see the Cicerone's Likelihood Livetime and Exposure: Livetime cubes). Exposure MapsThe likelihood consists of two factors: the first is dependent on the detected counts and differs between binned and unbinned likelihood calculations; and the second is equal to the exponential of the negative of the expected total number of counts Nexp for the source model. The exposure map is the total exposure—area multiplied by time—for a given position on the sky producing counts in the Region of Interest. Since the response function is a function of the photon energy, the exposure map is also a function of this energy. Thus, the counts produced by a source at a given position on the sky is the integral of the source flux and the exposure map (a function of energy) at that position. The exposure map is used for extended sources such as the diffuse Galactic and Extragalactic backgrounds, and not for individual sources. The exposure map should be computed over a Source Region that is larger than the Region of Interest by ~50%. This is necessary to insure that all source photons are included due to the size of the LAT instrument PSF at low energies.
|