Source code for imitation.regularization.regularizers

"""Implements the regularizer base class and some standard regularizers."""

import abc
from typing import Generic, Optional, Protocol, Type, TypeVar, Union

import numpy as np
import torch as th
from torch import optim

from imitation.regularization import updaters
from imitation.util import logger as imit_logger

# this is not actually a scalar, dimension check is still required for tensor.
Scalar = Union[th.Tensor, float]

R = TypeVar("R")
Self = TypeVar("Self", bound="Regularizer")
T_Regularizer_co = TypeVar(  # pytype: disable=not-supported-yet
    "T_Regularizer_co",
    covariant=True,
)


[docs]class RegularizerFactory(Protocol[T_Regularizer_co]): """Protocol for functions that create regularizers. The regularizer factory is meant to be used as a way to create a regularizer in two steps. First, the end-user creates a regularizer factory by calling the `.create()` method of a regularizer class. This allows specifying all the relevant configuration to the regularization algorithm. Then, the network algorithm finishes setting up the optimizer and logger, and calls the regularizer factory to create the regularizer. This two-step process separates the configuration of the regularization algorithm from additional "operational" parameters. This is useful because it solves two problems: #. The end-user does not have access to the optimizer and logger when configuring the regularization algorithm. #. Validation of the configuration is done outside the network constructor. It also allows re-using the same regularizer factory for multiple networks. """ def __call__( self, *, optimizer: optim.Optimizer, logger: imit_logger.HierarchicalLogger, ) -> T_Regularizer_co: """Constructs a regularizer from the factory. Args: optimizer: The optimizer used by the network. logger: The logger used by the network. """
[docs]class Regularizer(abc.ABC, Generic[R]): """Abstract class for creating regularizers with a common interface.""" optimizer: optim.Optimizer lambda_: float lambda_updater: Optional[updaters.LambdaUpdater] logger: imit_logger.HierarchicalLogger val_split: Optional[float]
[docs] def __init__( self, optimizer: optim.Optimizer, initial_lambda: float, lambda_updater: Optional[updaters.LambdaUpdater], logger: imit_logger.HierarchicalLogger, val_split: Optional[float] = None, ) -> None: """Initialize the regularizer. Args: optimizer: The optimizer to which the regularizer is attached. initial_lambda: The initial value of the regularization parameter. lambda_updater: A callable object that takes in the current lambda and the train and val loss, and returns the new lambda. logger: The logger to which the regularizer will log its parameters. val_split: The fraction of the training data to use as validation data for the lambda updater. Can be none if no lambda updater is provided. Raises: ValueError: if no lambda updater (``lambda_updater``) is provided and the initial regularization strength (``initial_lambda``) is zero. ValueError: if a validation split (``val_split``) is provided but it's not a float in the (0, 1) interval. ValueError: if a lambda updater is provided but no validation split is provided. ValueError: if a validation split is set, but no lambda updater is provided. """ if lambda_updater is None and np.allclose(initial_lambda, 0.0): raise ValueError( "If you do not pass a regularizer parameter updater your " "regularization strength must be non-zero, as this would " "result in no regularization.", ) if val_split is not None and ( not isinstance(val_split, float) or np.allclose(val_split, 0.0) or val_split <= 0 or val_split >= 1 ): raise ValueError( f"val_split = {val_split} must be a float strictly between 0 and 1.", ) if lambda_updater is not None and val_split is None: raise ValueError( "If you pass a regularizer parameter updater, you must also " "specify a validation split. Otherwise the updater won't have any " "validation data to use for updating.", ) elif lambda_updater is None and val_split is not None: raise ValueError( "If you pass a validation split, you must also " "pass a regularizer parameter updater. Otherwise you are wasting" " data into the validation split that will not be used.", ) self.optimizer = optimizer self.lambda_ = initial_lambda self.lambda_updater = lambda_updater self.logger = logger self.val_split = val_split self.logger.record("regularization_lambda", self.lambda_)
[docs] @classmethod def create( cls: Type[Self], initial_lambda: float, lambda_updater: Optional[updaters.LambdaUpdater] = None, val_split: float = 0.0, **kwargs, ) -> RegularizerFactory[Self]: """Create a regularizer.""" def factory( *, optimizer: optim.Optimizer, logger: imit_logger.HierarchicalLogger, ) -> Self: return cls( initial_lambda=initial_lambda, optimizer=optimizer, lambda_updater=lambda_updater, logger=logger, val_split=val_split, **kwargs, ) return factory
[docs] @abc.abstractmethod def regularize_and_backward(self, loss: th.Tensor) -> R: """Abstract method for performing the regularization step. The return type is a generic and the specific implementation must describe the meaning of the return type. This step will also call `loss.backward()` for the user. This is because the regularizer may require the loss to be called before or after the regularization step. Leaving this to the user would force them to make their implementation dependent on the regularizer algorithm used, which is prone to errors. Args: loss: The loss to regularize. """
[docs] def update_params(self, train_loss: Scalar, val_loss: Scalar) -> None: """Update the regularization parameter. This method calls the lambda_updater to update the regularization parameter, and assigns the new value to `self.lambda_`. Then logs the new value using the provided logger. Args: train_loss: The loss on the training set. val_loss: The loss on the validation set. """ if self.lambda_updater is not None: self.lambda_ = self.lambda_updater(self.lambda_, train_loss, val_loss) self.logger.record("regularization_lambda", self.lambda_)
[docs]class LossRegularizer(Regularizer[Scalar]): """Abstract base class for regularizers that add a loss term to the loss function. Requires the user to implement the _loss_penalty method. """ @abc.abstractmethod def _loss_penalty(self, loss: Scalar) -> Scalar: """Implement this method to add a loss term to the loss function. This method should return the term to be added to the loss function, not the regularized loss itself. Args: loss: The loss function to which the regularization term is added. """
[docs] def regularize_and_backward(self, loss: th.Tensor) -> Scalar: """Add the regularization term to the loss and compute gradients. Args: loss: The loss to regularize. Returns: The regularized loss. """ regularized_loss = th.add(loss, self._loss_penalty(loss)) regularized_loss.backward() self.logger.record("regularized_loss", regularized_loss.item()) return regularized_loss
[docs]class WeightRegularizer(Regularizer): """Abstract base class for regularizers that regularize the weights of a network. Requires the user to implement the _weight_penalty method. """ @abc.abstractmethod def _weight_penalty(self, weight: th.Tensor, group: dict) -> Scalar: """Implement this method to regularize the weights of the network. This method should return the regularization term to be added to the weight, not the regularized weight itself. Args: weight: The weight (network parameter) to regularize. group: The group of parameters to which the weight belongs. """
[docs] def regularize_and_backward(self, loss: th.Tensor) -> None: """Regularize the weights of the network, and call ``loss.backward()``.""" loss.backward() for group in self.optimizer.param_groups: for param in group["params"]: param.data = th.add(param.data, self._weight_penalty(param, group))
[docs]class LpRegularizer(LossRegularizer): """Applies Lp regularization to a loss function.""" p: int
[docs] def __init__( self, optimizer: optim.Optimizer, initial_lambda: float, lambda_updater: Optional[updaters.LambdaUpdater], logger: imit_logger.HierarchicalLogger, p: int, val_split: Optional[float] = None, ) -> None: """Initialize the regularizer.""" super().__init__(optimizer, initial_lambda, lambda_updater, logger, val_split) if not isinstance(p, int) or p < 1: raise ValueError("p must be a positive integer") self.p = p
def _loss_penalty(self, loss: Scalar) -> Scalar: """Returns the loss penalty. Calculates the p-th power of the Lp norm of the weights in the optimizer, and returns a scaled version of it as the penalty. Args: loss: The loss to regularize. Returns: The scaled pth power of the Lp norm of the network weights. """ del loss penalty = 0 for group in self.optimizer.param_groups: for param in group["params"]: penalty += th.linalg.vector_norm(param, ord=self.p).pow(self.p) return self.lambda_ * penalty
[docs]class WeightDecayRegularizer(WeightRegularizer): """Applies weight decay to a loss function.""" def _weight_penalty(self, weight, group) -> Scalar: """Returns the weight penalty. Args: weight: The weight to regularize. group: The group of parameters to which the weight belongs. Returns: The weight penalty (to add to the current value of the weight) """ return -self.lambda_ * group["lr"] * weight.data