Skip to content

olm.nn.activations.liglu

Classes

LiGLU(*args, **kwargs) LiGLU activation function.

class olm.nn.activations.liglu.ActivationBase(*args: Any, **kwargs: Any)

Bases: Module, ABC

Abstract base class for all activation functions.

Ensures a consistent interface for activation layers, handling device and dtype initialization. Subclasses must implement the forward method.

device

The device the module is on.

  • Type: torch.device, optional

dtype

The data type of the module parameters.

  • Type: torch.dtype

abstractmethod forward(x: torch.Tensor) → torch.Tensor

Apply activation to x.

class olm.nn.activations.liglu.LiGLU(*args: Any, **kwargs: Any)

Bases: ActivationBase

LiGLU activation function.

Implements the LiGLU variant (Linear GLU). LiGLU(x, W, V) = (xW) * (xV) Here: LiGLU(x) = gate * value (No activation on gate)

  • Parameters:
  • device (torch.device , optional) – Target device.
  • dtype (torch.dtype , optional) – Target data type.

forward(x: torch.Tensor) → torch.Tensor

Forward pass of LiGLU.

  • Parameters: x (torch.Tensor) – Input tensor.
  • Returns: Output tensor with half the last dimension.
  • Return type: torch.Tensor