Skip to content

Backend Module Design

Since v0.9, GPJax is built upon Flax's NNX module. This transition allows for more efficient parameter handling, improved integration with Flax and Flax-based libraries, and enhanced flexibility in model design. This notebook provides a high-level overview of the backend module design in GPJax. For an introduction to NNX, please refer to the official documentation.

import typing as tp

from flax import nnx

# Enable Float64 for more stable matrix inversions.
from jax import (
    config,
    grad,
)
import jax.numpy as jnp
import jax.tree_util as jtu
from jaxtyping import (
    Float,
    Num,
    install_import_hook,
)
import matplotlib as mpl
import matplotlib.pyplot as plt

from examples.utils import use_mpl_style
from gpjax.mean_functions import (
    AbstractMeanFunction,
    Constant,
)
from gpjax.parameters import (
    DEFAULT_BIJECTION,
    Parameter,
    PositiveReal,
    Real,
    transform,
)
from gpjax.typing import (
    Array,
    ScalarFloat,
)

config.update("jax_enable_x64", True)


with install_import_hook("gpjax", "beartype.beartype"):
    import gpjax as gpx


# set the default style for plotting
use_mpl_style()

cols = mpl.rcParams["axes.prop_cycle"].by_key()["color"]

Parameters

The biggest change bought about by the transition to an NNX backend is the increased support we now provide for handling parameters. As discussed in our Sharp Bits - Bijectors Doc, GPJax uses bijectors to transform constrained parameters to unconstrained parameters during optimisation. You may now register the support of a parameter using our Parameter class. To see this, consider the constant mean function who contains a single constant parameter whose value ordinarily exists on the real line. We can register this parameter as follows:

constant_param = Parameter(value=1.0, tag=None)
meanf = Constant(constant_param)
print(meanf)
Constant(
  constant=Parameter(
    value=Array(1., dtype=float64, weak_type=True),
    _tag=None
  )
)

However, suppose you wish your mean function's constant parameter to be strictly positive. This is easy to achieve by using the correct Parameter type which, in this case, will be the PositiveReal. However, any Parameter that subclasses from Parameter will be transformed by GPJax.

issubclass(PositiveReal, Parameter)
True

Injecting this newly constrained parameter into our mean function is then identical to before.

constant_param = PositiveReal(value=1.0)
meanf = Constant(constant_param)
print(meanf)
Constant(
  constant=PositiveReal(
    value=Array(1., dtype=float64, weak_type=True),
    _tag='positive'
  )
)

Were we to try and instantiate the PositiveReal class with a negative value, then an explicit error would be raised.

try:
    PositiveReal(value=-1.0)
except ValueError as e:
    print(e)
Expected parameter value to be strictly positive. Got -1.0.

Parameter Transforms

With a parameter instantiated, you likely wish to transform the parameter's value from its constrained support onto the entire real line. To do this, you can apply the transform function to the parameter. To control the bijector used to transform the parameter, you may pass a set of bijectors into the transform function. Under-the-hood, the transform function is looking up the bijector of a parameter using it's _tag field in the bijector dictionary, and then applying the bijector to the parameter's value using a tree map operation.

print(constant_param._tag)
positive

For most users, you will not need to worry about this as we provide a set of default bijectors that are defined for all the parameter types we support. However, see our Kernel Guide Notebook to see how you can define your own bijectors and parameter types.

print(DEFAULT_BIJECTION[constant_param._tag])
tfp.bijectors.Softplus("softplus", batch_shape=[], min_event_ndims=0)

We see here that the Softplus bijector is specified as the default for strictly positive parameters. To apply this, we must first realise the state of our model. This is achieved using the split function provided by nnx.

_, _params = nnx.split(meanf, Parameter)

tranformed_params = transform(_params, DEFAULT_BIJECTION, inverse=True)

The parameter's value was changed here from 1. to 0.54132485. This is the result of applying the Softplus bijector to the parameter's value and projecting its value onto the real line. Were the parameter's value to be closer to 0, then the transformation would be more pronounced.

_, _close_to_zero_state = nnx.split(Constant(PositiveReal(value=1e-6)), Parameter)

transform(_close_to_zero_state, DEFAULT_BIJECTION, inverse=True)
State({
  'constant': VariableState(
    type=PositiveReal,
    value=Array(-13.81551006, dtype=float64),
    _tag='positive'
  )
})

Transforming Multiple Parameters

In the above, we transformed a single parameter. However, in practice your parameters may be nested within several functions e.g., a kernel function within a GP model. Fortunately, transforming several parameters is a simple operation that we here demonstrate for a conjugate GP posterior (see our Regression Notebook for detailed explanation of this model.).

kernel = gpx.kernels.Matern32()
meanf = gpx.mean_functions.Constant()

prior = gpx.gps.Prior(mean_function=meanf, kernel=kernel)

likelihood = gpx.likelihoods.Gaussian(100)
posterior = likelihood * prior
print(posterior)
ConjugatePosterior(
  prior=Prior(
    kernel=Matern32(
      active_dims=slice(None, None, None),
      n_dims=None,
      compute_engine=<gpjax.kernels.computations.dense.DenseKernelComputation object at 0x7f0f32178580>,
      lengthscale=PositiveReal(
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      ),
      variance=PositiveReal(
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      )
    ),
    mean_function=Constant(
      constant=Real(
        value=Array(0., dtype=float64, weak_type=True),
        _tag='real'
      )
    ),
    jitter=1e-06
  ),
  likelihood=Gaussian(
    obs_stddev=PositiveReal(
      value=Array(1., dtype=float64, weak_type=True),
      _tag='positive'
    ),
    num_datapoints=100,
    integrator=<gpjax.integrators.AnalyticalGaussianIntegrator object at 0x7f0f321d2920>
  ),
  jitter=1e-06
)

Now contained within the posterior PyGraph here there are four parameters: the kernel's lengthscale and variance, the noise variance of the likelihood, and the constant of the mean function. Using NNX, we may realise these parameters through the nnx.split function. The split function deomposes a PyGraph into a GraphDef and State object. As the name suggests, State contains information on the parameters' state, whilst GraphDef contains the information required to reconstruct a PyGraph from a give State.

graphdef, state = nnx.split(posterior)
print(state)
State({
  'likelihood': {
    'obs_stddev': VariableState(
      type=PositiveReal,
      value=Array(1., dtype=float64, weak_type=True),
      _tag='positive'
    )
  },
  'prior': {
    'kernel': {
      'lengthscale': VariableState(
        type=PositiveReal,
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      ),
      'variance': VariableState(
        type=PositiveReal,
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      )
    },
    'mean_function': {
      'constant': VariableState(
        type=Real,
        value=Array(0., dtype=float64, weak_type=True),
        _tag='real'
      )
    }
  }
})

The State object behaves just like a PyTree and, consequently, we may use JAX's tree_map function to alter the values of the State. The updated State can then be used to reconstruct our posterior. In the below, we simply increment each parameter's value by 1.

updated_state = jtu.tree_map(lambda x: x + 1, state)
print(updated_state)
State({
  'likelihood': {
    'obs_stddev': VariableState(
      type=PositiveReal,
      value=Array(2., dtype=float64, weak_type=True),
      _tag='positive'
    )
  },
  'prior': {
    'kernel': {
      'lengthscale': VariableState(
        type=PositiveReal,
        value=Array(2., dtype=float64, weak_type=True),
        _tag='positive'
      ),
      'variance': VariableState(
        type=PositiveReal,
        value=Array(2., dtype=float64, weak_type=True),
        _tag='positive'
      )
    },
    'mean_function': {
      'constant': VariableState(
        type=Real,
        value=Array(1., dtype=float64, weak_type=True),
        _tag='real'
      )
    }
  }
})

Let us now use NNX's merge function to reconstruct the posterior distribution using the updated state.

updated_posterior = nnx.merge(graphdef, updated_state)
print(updated_posterior)
ConjugatePosterior(
  jitter=1e-06,
  likelihood=Gaussian(
    integrator=<gpjax.integrators.AnalyticalGaussianIntegrator object at 0x7f0f321d2920>,
    num_datapoints=100,
    obs_stddev=PositiveReal(
      _tag='positive',
      value=Array(2., dtype=float64, weak_type=True)
    )
  ),
  prior=Prior(
    jitter=1e-06,
    kernel=Matern32(
      active_dims=slice(None, None, None),
      compute_engine=<gpjax.kernels.computations.dense.DenseKernelComputation object at 0x7f0f32178580>,
      lengthscale=PositiveReal(
        _tag='positive',
        value=Array(2., dtype=float64, weak_type=True)
      ),
      n_dims=None,
      variance=PositiveReal(
        _tag='positive',
        value=Array(2., dtype=float64, weak_type=True)
      )
    ),
    mean_function=Constant(
      constant=Real(
        _tag='real',
        value=Array(1., dtype=float64, weak_type=True)
      )
    )
  )
)

However, we begun this point of conversation with bijectors in mind, so let us now see how bijectors may be applied to a collection of parameters in GPJax. Fortunately, this is very straightforward, and we may simply use the transform function as before.

transformed_state = transform(state, DEFAULT_BIJECTION, inverse=True)
print(transformed_state)
State({
  'likelihood': {
    'obs_stddev': VariableState(
      type=PositiveReal,
      value=Array(0.54132485, dtype=float64),
      _tag='positive'
    )
  },
  'prior': {
    'kernel': {
      'lengthscale': VariableState(
        type=PositiveReal,
        value=Array(0.54132485, dtype=float64),
        _tag='positive'
      ),
      'variance': VariableState(
        type=PositiveReal,
        value=Array(0.54132485, dtype=float64),
        _tag='positive'
      )
    },
    'mean_function': {
      'constant': VariableState(
        type=Real,
        value=Array(0., dtype=float64, weak_type=True),
        _tag='real'
      )
    }
  }
})

We may also (re-)constrain the parameters' values by setting the inverse argument of transform to False.

retransformed_state = transform(transformed_state, DEFAULT_BIJECTION, inverse=False)

Fine-Scale Control

One of the advantages of being able to split and re-merge the PyGraph is that we are able to gain fine-scale control over the parameters' whose state we wish to realise. This is by virtue of the fact that each of our parameters now inherit from gpjax.parameters.Parameter. In the former, we were simply extracting any Parametersubclass from the posterior. However, suppose we only wish to extract those parameters whose support is the positive real line. This is easily achieved by altering the way in which we invoke nnx.split.

graphdef, positive_reals, other_params = nnx.split(posterior, PositiveReal, ...)
print(positive_reals)
State({
  'likelihood': {
    'obs_stddev': VariableState(
      type=PositiveReal,
      value=Array(1., dtype=float64, weak_type=True),
      _tag='positive'
    )
  },
  'prior': {
    'kernel': {
      'lengthscale': VariableState(
        type=PositiveReal,
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      ),
      'variance': VariableState(
        type=PositiveReal,
        value=Array(1., dtype=float64, weak_type=True),
        _tag='positive'
      )
    }
  }
})

Now we see that we have two state objects: one containing the positive real parameters and the other containing the remaining parameters. This functionality is exceptionally useful as it allows us to efficiently operate on a subset of the parameters whilst leaving the others untouched. Looking forward, we hope to use this functionality in our Variational Inference Approximations to perform more efficient updates of the variational parameters and then the model's hyperparameters.

NNX Modules

To conclude this notebook, we will now demonstrate the ease of use and flexibility offered by NNX modules. To do this, we will implement a linear mean function using the existing abstractions in GPJax.

For inputs \(x_n \in \mathbb{R}^d\), the linear mean function \(m(x): \mathbb{R}^d \to \mathbb{R}\) is defined as: $$ m(x) = \alpha + \sum_{i=1}^d \beta_i x_i $$ where \(\alpha \in \mathbb{R}\) and \(\beta_i \in \mathbb{R}\) are the parameters of the mean function. Let's now implement that using the new NNX backend.

class LinearMeanFunction(AbstractMeanFunction):
    def __init__(
        self,
        intercept: tp.Union[ScalarFloat, Float[Array, " O"], Parameter] = 0.0,
        slope: tp.Union[ScalarFloat, Float[Array, " D O"], Parameter] = 0.0,
    ):
        if isinstance(intercept, Parameter):
            self.intercept = intercept
        else:
            self.intercept = Real(jnp.array(intercept))

        if isinstance(slope, Parameter):
            self.slope = slope
        else:
            self.slope = Real(jnp.array(slope))

    def __call__(self, x: Num[Array, "N D"]) -> Float[Array, "N O"]:
        return self.intercept.value + jnp.dot(x, self.slope.value)

As we can see, the implementation is straightforward and concise. The AbstractMeanFunction module is a subclass of nnx.Module and may, therefore, be used in any split or merge call. Further, we have registered the intercept and slope parameters as Real parameter types. This registers their value in the PyGraph and means that they will be part of any operation applied to the PyGraph e.g., transforming and differentiation.

To check our implementation worked, let's now plot the value of our mean function for a linearly spaced set of inputs.

N = 100
X = jnp.linspace(-5.0, 5.0, N)[:, None]

meanf = LinearMeanFunction(intercept=1.0, slope=2.0)
plt.plot(X, meanf(X))
[<matplotlib.lines.Line2D at 0x7f0f2e504a00>]

png

Looks good! To conclude this section, let's now parameterise a GP with our new mean function and see how gradients may be computed.

y = jnp.sin(X)
D = gpx.Dataset(X, y)

prior = gpx.gps.Prior(mean_function=meanf, kernel=gpx.kernels.Matern32())
likelihood = gpx.likelihoods.Gaussian(D.n)
posterior = likelihood * prior

We'll compute derivatives of the conjugate marginal log-likelihood, with respect to the unconstrained state of the kernel, mean function, and likelihood parameters.

graphdef, params, others = nnx.split(posterior, Parameter, ...)
params = transform(params, DEFAULT_BIJECTION, inverse=True)


def loss_fn(params: nnx.State, data: gpx.Dataset) -> ScalarFloat:
    params = transform(params, DEFAULT_BIJECTION)
    model = nnx.merge(graphdef, params, *others)
    return -gpx.objectives.conjugate_mll(model, data)


param_grads = grad(loss_fn)(params, D)

In practice, you would wish to perform multiple iterations of gradient descent to learn the optimal parameter values. However, for the purposes of illustration, we use another tree_map in the below to update the parameters' state using their previously computed gradients. As you can see, the really beauty in having access to the model's state is that we have full control over the operations that we perform to the state.

LEARNING_RATE = 0.01
optimised_params = jtu.tree_map(
    lambda _params, _grads: _params + LEARNING_RATE * _grads, params, param_grads
)

Now we will plot the updated mean function alongside its initial form. To achieve this, we first merge the state back into the model using merge, and we then simply invoke the model as normal.

optimised_posterior = nnx.merge(graphdef, optimised_params, *others)

fig, ax = plt.subplots()
ax.plot(X, optimised_posterior.prior.mean_function(X), label="Updated mean function")
ax.plot(X, meanf(X), label="Initial mean function")
ax.legend()
ax.set(xlabel="x", ylabel="m(x)")
[Text(0.5, 0, 'x'), Text(0, 0.5, 'm(x)')]

png

Conclusions

In this notebook we have explored how GPJax's Flax-based backend may be easily manipulated and extended. For a more applied look at this, see how we construct a kernel on polar coordinates in our Kernel Guide notebook.

System configuration

%reload_ext watermark
%watermark -n -u -v -iv -w -a 'Thomas Pinder'
Author: Thomas Pinder

Last updated: Tue Nov 26 2024

Python implementation: CPython
Python version       : 3.10.15
IPython version      : 8.29.0

gpjax     : 0.9.3
flax      : 0.10.2
jax       : 0.4.27
matplotlib: 3.9.2
jaxtyping : 0.2.36

Watermark: 2.5.0