Chapter 5
Inference Methodology

Highly multivariate data is often challenging to model due to the “curse of dimensionality” Bellman (1957). Much of the work in this thesis is concerned with reducing the computational burden associated with modelling the response of vegetation to climate using the modern training data (see Section 1.1.1).

This is achieved through separate analysis of the marginal responses of individual plant taxa to climate. The approach necessarily ignores between taxa dependencies but allows for reduction of overall computation. The separate, marginal analyses can then be brought together post-hoc. This is referred to as the inference-via-marginals posterior and is an approximation to the joint posterior of the full model. Situations where the approximation is poor and where it is excellent, or even exact, are identified.

Details related to working on a discrete grid across the location space are also investigated. This chapter serves as a review and assessment of the preceding chapters methodology. The methods are brought together in preparation for the application to the pollen dataset in Chapter 6.

In particular, the concern here is with inference on multiple latent spatial Gaussian processes defined on a lattice. The INLA methodology introduced in Section 4.1 is unsuitable for this task as it can only deal with one such spatial process at a time (see Section 5.1.4). If the model does not disjoint-decompose (Section 3.2) exactly then approximations to the model that do decompose must be sought. The accuracy of these approximations must then be tested, as per Section 5.3.

The goal is inference on latent random variables X, given counts data Y defined at discrete locations C. X is composed of NT processes which are Gaussian (or approximately so) given the data. Thus at each value of C, there are NT counts or proportions arising from NT potentially dependent X values. As the posterior is GMRF due to the methods introduced in Section 4.1, it is expressed via the posterior mean vector μ and precision matrix Q. i.e. the posterior distribution of X is given approximately by:

      C
π!X   M  π!YjSXj      GM  RF  !XSμ, Q
     j  1
(5.1)

The goal is therefore inference on the μ and Q terms. By disjoint-decomposition into marginals defined as the product of independent multivariate Xi,i 1,,NT across locations C becomes:

                   NT           NT
GM  RF  !XSμ,Q       M π!XiSYi      M  GM  RF  !XiSμi,Qi
                   i  1          i  1
(5.2)

The INLA method then delivers the terms μi and Qi that define the independent spatial Xi given the data. The error incurred due to this decomposition is the subject of Section 5.3.

5.1 Reasons for Disjoint-Decomposition

5.1.1 Parallelisation

If the overall problem disjoint-decomposes (see Section 3.2) then the exploitation of parallel programming resources is trivial. Inference on each disjoint module may be performed entirely independently of the others and thus may be done at the same time on multiple processors. No specialist code or hardware is required and joint summary statistics are simple and quick calculations that are made on a single processor.

5.1.2 Memory Usage

When responses are modelled non-parametrically (see Section 1.1.1) as for the palaeoclimate dataset, the number of parameters is of the order of 103 for a two dimensional climate space. Ultimately, the goal is to work with at least three climate variables (leading to the order of millions of parameters) and also to integrate out the unknown hyperparameters. The memory required simply to store such a burdensome model may quickly overwhelm even modern, high specification personal computers.

Memory usage is therefore one of the most difficult issues associated with performing inference on this dataset. It is crucial to reduce both the computational overhead and the size the model takes up in memory. Breaking the model down into a series of approximately disjoint / conditionally independent marginals negates the need to store and manipulate all of the parameters at once on a single machine.

5.1.3 Inverse Problem

Another large saving in terms of computation is in the inversion of the forward model; a high dimensional integral (to find the joint marginal likelihood) is replaced with a product of uni-dimensional integrals. These may be evaluated without resorting to Monte-Carlo based sampling methods that are required for the estimation of multidimensional integrals.

5.1.4 Compatibility with the INLA Method

If the joint model does not decompose, then inference must be performed for the entire model at once. Interaction terms between marginals are non-zero and must be modelled explicitly. For problems consisting of multiple smooth surfaces giving rise to vectors of counts, such as the motivating palaeoclimate problem, this results in difficulties for the INLA inference method.

Interaction may occur at one of two levels in the model hierarchy; either in the prior or the likelihood. The former models smooth surfaces that are not independent across locations; the latter models independent smooth surfaces that jointly give rise to non-independent counts. This is a continuation of the discussion in Section 3.2.2.

Prior and Hyperparameters

If the interactions between marginals are modelled at the latent variable stage, then the joint prior must contain these terms. Specifically, the joint prior precision matrix must contain non-zero terms for the interactions. If they are not known a-priori then unknown hyperparameters must be introduced to model these inter-process precisions.

The INLA method deals with hyperparameters via numerical integration. The entire vector of hyperparameters is set on a discrete grid, as per Section 4.1.3. This approach requires that the number of hyperparameters is low; any more than five or so and the method runs into computational difficulty. The motivating palaeoclimate problem has 28 plant taxa; even crude modelling of the interactions with a single hyperparameter governing each taxon-taxon interaction results in 378 additional hyperparameters; this is far more than the INLA method can cope with (Rue et al. (2008)).

Likelihood and Taylor Expansions

Modelling the interactions at the data level eliminates the need for additional interaction hyperparameters. Interaction terms are placed in the likelihood precision matrix and are thus parameters rather than hyperparameters; however the data are not conditionally independent. The Taylor series expansions in the GMRF approximation are no longer univariate, resulting in a massive increase in the computation required to fit the approximation (see Section 4.1.2). This is a fundamental challenge to the existing INLA methodology.

5.2 Multivariate Normal Model

Much of this chapter focusses on the marginal analysis of data in a Gaussian setting. The reasons for exploration of such a problem in a purely Gaussian framework are as follows:

  1. For simple exploration and illustration of key points; the availability of the posterior in closed form is the primary motivation.
  2. For synthesis with the Gaussian approximation technique described in Section 4.1. This technique applies a Gaussian prior and approximates the posterior with a Gaussian.

Given a multivariate normal prior π!X and multivariate normal likelihood π!Y SX, the posterior π!XSY is multivariate normal with mean and precision matrix given by

μ      !QX     QY     1!QX μX    QY  μY

Q      QX     QY                                        (5.3)

where the prior precision matrix is QX and the likelihood precision matrix is QY .

5.2.1 Conditions for Perfect Disjoint-Decomposition

If the joint posterior expresses zero precision between processes then a fully joint inference may be done exactly via the marginals. There are in fact two situations in which this occurs:

  1. The underlying processes are truly independent of each other
  2. The joint model creates a posterior with conditional independence across the latent variables, regardless of any dependency structure suggested by the data 1

The second situation for a perfect inference-via-marginals is clearly illustrated with two common examples. These are discussed in Section 5.2.2 and Section 5.2.

The terminology used in this thesis will be that the joint model disjoint-decomposes exactly (see Section 3.2).

It can immediately be seen from Equation (5.3) that the condition for exact disjoint-decomposition of the posterior is that both the prior and the likelihood disjoint-decompose. The form of the precision matrix shows whether a density will decompose; if it is block-diagonal, then the blocks each represent a disjoint part of the full model. See Table 5.1 for illustration.


Table 5.1: Joint prior precision matrix entries, using the indexing system i1,j1i2,j2 means precision between processes i1 and i2 at locations j1 and j2, respectively (processes here are the latent fields). Intra-process precision is highlighted in yellow and inter-process precision is unhighlighted. If the processes are conditionally independent given the locations then all values in the non-highlighted sub-matrices will be zero in the prior precision matrix, so that it is block-diagonal. This table shows the precision matrix for number of processes NT 4 and number of discrete locations NL 4.
















1,11,11,11,21,11,31,11,41,12,11,12,21,12,31,12,41,13,11,13,21,13,31,13,41,14,11,14,21,14,31,14,4
















1,21,1 1,21,21,21,31,21,41,22,11,22,21,22,31,22,41,23,11,23,21,23,31,23,41,24,11,24,21,24,31,24,4
















1,31,1 1,31,21,31,31,31,41,32,11,32,21,32,31,32,41,33,11,33,21,33,31,33,41,34,11,34,21,34,31,34,4
















1,41,1 1,41,21,41,31,41,41,42,11,42,21,42,31,42,41,43,11,43,21,43,31,43,41,44,11,44,21,44,31,44,4
















2,11,1 2,11,22,11,32,11,42,12,12,12,22,12,32,12,42,13,12,13,22,13,32,13,42,14,12,14,22,14,32,14,4
















2,21,1 2,21,22,21,32,21,42,22,12,22,22,22,32,22,42,23,12,23,22,23,32,23,42,24,12,24,22,24,32,24,4
















2,31,1 2,31,22,31,32,31,42,32,12,32,22,32,32,32,42,33,12,33,22,33,32,33,42,34,12,34,22,34,32,34,4
















2,41,1 2,41,22,41,32,41,42,42,12,42,22,42,32,42,42,43,12,43,22,43,32,43,42,44,12,44,22,44,32,44,4
















3,11,1 3,11,23,11,33,11,43,12,13,12,23,12,33,12,43,13,13,13,23,13,33,13,43,14,13,14,23,14,33,14,4
















3,21,1 3,21,23,21,33,21,43,22,13,22,23,22,33,22,43,23,13,23,23,23,33,23,43,24,13,24,23,24,33,24,4
















3,31,1 3,31,23,31,33,31,43,32,13,32,23,32,33,32,43,33,13,33,23,33,33,33,43,34,13,34,23,34,33,34,4
















3,41,1 3,41,23,41,33,41,43,42,13,42,23,42,33,42,43,43,13,43,23,43,33,43,43,44,13,44,23,44,33,44,4
















4,11,1 4,11,24,11,34,11,44,12,14,12,24,12,34,12,44,13,14,13,24,13,34,13,44,14,14,14,24,14,34,14,4
















4,21,1 4,21,24,21,34,21,44,22,14,22,24,22,34,22,44,23,14,23,24,23,34,23,44,24,14,24,24,24,34,24,4
















4,31,1 4,31,24,31,34,31,44,32,14,32,24,32,34,32,44,33,14,33,24,33,34,33,44,34,14,34,24,34,34,34,4
















4,41,1 4,41,24,41,34,41,44,42,14,42,24,42,34,42,44,43,14,43,24,43,34,43,44,44,14,44,24,44,34,44,4

















The assumption that the precision of process i1 at location j1 given process i2 at location j2 ( j1 is zero is logical; there will not be interaction between a plant taxon at one climate location and a different taxon at another, disparate location in climate space. This results in a banded overall precision matrix and is similar to the Separable Models described in Finkenstdt et al. (2006). However, their approach cannot be exploited here. Separable models involve modelling two separate precision matrices; typically a spatial matrix and a temporal matrix. The spatio-temporal precision matrix is then taken as the Kronecker product of these two matrices. Taking a Kronecker product of an intra-process precision matrix and an inter-process precision matrix would impose two modelling aspects that are unsuitable for this work. Specifically, (i) implementation of a separable model would require using a common spatial precision matrix for all processes and (ii) would lead to non-zero i1,j1i2,j2 terms in the precision matrices for i1 ( i2 and j1 ( j2.

5.2.2 Compositional Independence

Although it would seem that compositional data analysis must necessarily model interaction between the components of the composition (in the pollen dataset, the components are the various plant taxa), Aitchison demonstrates that any statistical analysis making use of the Dirichlet distribution is in fact imposing a strong implied independence structure (Aitchison (1986) 3.4). Similarly, any logistic-normal (Section 3.5.4) distribution with diagonal precision matrix will impose the same strong implied independence structure.

For example, in a Bayesian setting, joint inference done on the latent vector of probability parameters P governing a compositional counts vector Y might proceed as follows:
Likelihood is Multinomial:

            ---n!---NT  yi
π!Y ;n,P        NT    M  pi
            L i  1 yi!i 1
(5.4)

where n PiY i
Prior is Dirichlet

                N
          --1--- T  αi  1
π!P ;α      B! α   Mi  1 pi
(5.5)

where α is a vector of hyperparameters
The posterior is then Dirichlet due to conjugacy

              ----1-----NT  αi  yi 1
π!P ;α   Y       B!α    Y   M  p i
                        i  1
(5.6)

The vector P is subject to the constraint that PiPi 1 and the vector of counts Y is subject to the constraint that PiY i n.

However, the Multinomial and the Dirichlet actually enforce conditional independence given the constraints.

The Multinomial can be expressed as a product of Poisson distributions, with parameters equal to nP, conditioned on the sum being equal to the total count n. This sum (Piyi n) itself follows a Poisson distribution with rate parameter n.

              NT
π!Y ;n,P      L i--1 Poisson!yi;λi-
               Poisson!n; n
(5.7)

with λi n pi and Poisson!yi;λi y
λii e--λ
  yi!

Similarly, a Dirichlet with parameter vector η may be expressed as a product of Gamma distributions, with shape parameters η and rate parameters all equal to Piηi, conditioned on the sum = 1 following a Gamma distribution with shape and rate both equal to Piηi.

          LNTi- 1-Gamma!pi;-ηi,Pk-ηk---
π!P ;η       Gamma!1;      η ,   η
                      Pk  k Pk  k
(5.8)

Therefore, to perform joint inference given a vector of counts and using a Dirichlet prior and a Multinomial likelihood, no accuracy is lost in performing marginal inferences on each part of the composition and then conditioning on the sum. This result is regardless of the value of the parameters of the distributions.

If sampling from the Dirichlet posterior is required, this can be achieved by sampling from the Gamma marginals and then rescaling such that the sum is one. In fact, this is the usual algorithm for sampling from a Dirichlet distribution. The post-hoc conditioning or rescaling is the only step that requires joint knowledge of the marginals.

5.3 Sensitivity to Inference via Marginals

If the joint model disjoint-decomposes then the inference-via-marginals approximation is exact. If not, the approximation amounts to setting non-zero terms in the overall precision matrix to be zero. This is equivalent to breaking some links in the graph of the model, specifically the inter-process links.

The accuracy of the inference-via-marginals approximation depends on several factors. The magnitude of the non-zero terms that are set to zero to facilitate decomposition are the most obvious of these. The further these are from zero, the greater the interaction and hence the worse the inference-via-marginals approximation will be. However, given non-zero interactions, several other factors will impact the level of accuracy.

Ultimately, the interest is in the differences between a full joint model and the inference-via-marginals model in terms of the inverse problem. Of course, if the forward model disjoint-decomposes exactly, then the inverse predictive distributions will be identical for the two models.

The worst case scenario is presented in Figure 5.1. A single surface is replicated T times; random counts are generated at various points in the location space. Inverse cross-validation predictive distributions are formed and the joint predictive distribution is found by taking the product and normalising (see Section 3.2). However, this model treats the surfaces as independent; they are in fact fully correlated. This results in a linear increase with Δ, the percentage of points lying outside their 95% cross-validation highest inverse predictive density with T.

Thus the inference-via-marginals model is a poor approximation; correlation of 1 between the surfaces (as they are all identical) represents the upper bound of inaccuracy of the inference-via-marginals approximation. Even if the data are simulated for each replication of the surface independently, the model does not disjoint-decompose.

To see this, take s1!c s2!c both the same smooth function of c. Random samples x1 N!s1 and x2 N!s2 will have high corr!x1,x2 but low corr!x1,x2Sc. So inference on the forward model may be completed marginally. Thus the result in Figure 5.1 is, at first glance, counter-intuitive. Independent counts should yield a tighter predictive density on the correct location. Δ should converge to zero as the predictive cross-validation densities converge to Dirac distributions on the correct locations.

The reason why this is not so is a peculiarity of the inverse problem; the shape of the response surface in this case is such that for any given location, there is typically one or more other locations with the same (or a very similar) value for the response. When a point is left out, the posterior variance at that point is greater than the locations for which there is data (for illustration, see Figure 2.3(b)). Thus the marginal likelihood is higher for these other locations; in turn the inverse predictive density will give greater weight to those locations with data and a similar response to the correct location. For a single response surface, this does not cause a problem. It is when analysing multiple surfaces that the issue arises.

Although the correct, left-out location has non-zero predictive density for a single surface (and is within the bounds of the 95% highest posterior predictive density 95% of the time), it will tend to zero as T increases. Multiplication of the density with itself T times will cause the location with the single highest density to gain all the mass as the number of replications of the surface increases. Thus, the correct location loses predictive probability mass and Δ increases.

In the simplest example, suppose inference on a single surface leads to the (correct) location predictive density of 75% probability mass a location A and 25% at location B, with location B the correct location. This is ok in terms of the Δ statistic; B is within the 95% HPD region. Now suppose that we are presented with new data from the same response, but treat it as independent. We might find, again without error, that this data suggests P!A 70% and P!B 30%. Again, this is ok. Assumption that the data are independent however, leads to a multiplication and re-normalisation of these values so that P!A 87.5% and P!B 12.5%. A third dataset yields P!A 80% and P!B 20%. The resultant multiplicative joint predictive distribution then has P!A 96.55% and P!B 3.45%. Now B is outside the 95% HPD region.


PIC


Fig. 5.1: A single smooth surface gives rise to counts data at locations. Cross-validation in the inverse sense (location given count) is used to find Δ, the percentage of points falling outside their 95% highest predictive distribution region. As the data are simulated and the model fitted is the same as the model used to generate the data, Δ is about 5%.
Taking T replications of the surface and taking the normalised product of the cross-validation inverse predictive distributions is equivalent to fitting the inference-via-marginals model to a joint model that has maximum inter-process correlation: the more replications of the single surface, the worse the approximation and the higher the Δ statistic. The above graph depicts 10 replications with randomly generated data for each value of T (scatter plot). The line is a mean across these 10 replications.


This issue will not arise when there is no other region of the location space with a very similar counts data vector. Using the exact same model and code used to produce Figure 5.1, but generating a new random response surface for each replication results in a Δ statistic of exactly zero for 20 surfaces. This is because each surface carries independent information on the location given counts data. Figure 5.2 shows a plot of Δ!T for random independent surfaces.

Note that this convergence to a probability of one at the correct location will also occur for the case of replicated identical surfaces when the cross-validation is performed using the saturated posterior.


PIC


Fig. 5.2: A single smooth surface gives rise to counts data at locations. Cross-validation in the inverse sense (location given count) is used to find Δ, the percentage of points falling outside their 95% highest predictive distribution region.
Generating T such independent surfaces and taking the normalised product of the cross-validation inverse predictive distributions is equivalent to fitting the inference-via-marginals model to a joint model that has zero inter-process correlation. The inference-via-marginals approximation is exact and as the independent information increases, the inverse predictive densities converge towards Dirac distributions on the correct locations (see Section 5.3.1).


5.3.1 Discrete HPD Regions

The reason why Δ does not converge to 5% and instead converges to 0% as the number of conditionally independent counts increases is due to the discrete grid. As per Section 3.1.3, the HPD region contains 95% or more of the total probability mass. Therefore, the expected value of Δ is B 5% for each independent surface.

As more of these conditionally independent components are brought together, each with Δ B 5%, the predictive distributions become increasingly peaked. They are centred on the correct location (under the correct model) and so the probability mass becomes focused at the correct location. Eventually, the 95% HPD region becomes smaller than the grid spacing so that all the mass is concentrated on a single gridpoint. Using the algorithm in Section 3.1.3 for constructing discrete 95% HPD regions then results in selection of the single gridpoint that contains all this mass. As this is the correct location, none of the points lie outside their corresponding 95% (or more) HPD region and Δ tends to zero. A graphic illustration of this phenomenon is presented in Figure 5.3.


PIC
(a) 2 counts per location
PIC
(b) 20 counts per location

Fig. 5.3: UPPER PANELS: A linear response gives rise to Gaussian counts (blue circles). If there are n independent counts at each location, this is analagous to multiple conditionally independent counts. These counts data are used to inform the mean posterior response on a coarse grid (red line) and fine grid (green line) and posterior variances (the broken lines depict the mean response 2σ with σ the posterior standard deviation).
LOWER PANELS: As the number of independent observations increases, the inverse predictive distribution becomes sharper; a mean count of 5 given 2 counts (Figure (a)) gives rise to a broader predictive distribution than a mean count of 5 given 20 counts (Figure (b)).
The probability mass on the coarse grid is depicted without joining the points (red circles). The fine grid approximates continuous space and so a line is used.
Although the coarse and fine grids deliver similar results at the coarse gridpoints (subject to the difference in scale due to the differing number of evaluation points), Figure (b) shows how the 95% HPD region will differ. The coarse grid now concentrates all mass at a single point. Hence the discrete 95% or greater HPD region will actually converge to a 100% HPD region and the Δ statistic converges to 0%.


5.3.2 Nested Constrained Models

Nested constrained models represent an interesting opportunity to apply disjoint-decomposition to a joint model that does not decompose exactly as the product of its marginals (see Section 3.5.6).

Given knowledge of the nesting structure, the joint model may be exactly expressed as the product of the marginals across all levels of the nesting hierarchy, where the dependencies within levels are accounted for in the likelihood using knowledge of the counts and responses of higher levels in the structure.

The advantage of such models is that they will disjoint-decompose iff nesting structure is known a-priori. If the nesting is not known, the marginals model at the lowest level only will be a poor fit to the data. Figure 5.4 shows results for inference-via-marginals models fitted to the same data-sets used in Figure 3.14, but with the added task of inference of the forward model. It is clear that attempting to model the joint model as the product of the marginals without exploitation of the nesting structure results in a poor model fit. In contrast, knowledge of the nesting structure allows for exact disjoint-decomposition of the model.


PIC


Fig. 5.4: Sample density for the Δ fit statistic, which is 5% in theory, under the correct model. The correct model is nested with two levels. If the nesting structure is unknown then the decomposed model (red density) is a poor fit. The sample Δs were obtained via multiple random runs. The inference via-marginals model is a poor approximation to the joint model, resulting in many more points lying outside their 95% HPD predictive region.


5.4 Conclusions

Inference on multivariate models comprised of multiple spatial processes may be performed in disjoint modules; provided there is no interaction between these modules in either the prior or the likelihood. Decomposition is into the marginals for each process.

The multivariate normal setting delivers insight into more general models, where the posteriors may be approximated with a GMRF. This approximation requires that the model decomposes. If it does not, then use of a disjoint-decomposable model is equivalent to a model with the interaction terms set to zero. This is an approximation, the accuracy of which is determined by the degree of correlation between the marginals. When such correlation is non-zero, the loss in accuracy increases linearly with the number of non-independent processes. Accuracy here is measured by the percentage of points lying outside their 95% discrete HPD region under leave-one-out cross-validation inverse predictive distributions.

Compositional models, in which both the data and the likelihood parameters are constrained under summation, represent a class of model that do not decompose. However, many compositional models do decompose for inference on the model parameters, subject to a post-hoc conditioning or rescaling. Compositional models which do not disjoint-decompose may in do so given knowledge of a nesting structure. Such structures must be known a-priori to facilitate decomposition of the model.