SambaNova compilation overview

When you compile a PyTorch model with SambaFlow, the model is processed at several layers. This document explains what happens at each layer, and discusses arguments you can use to affect processing at each layer.

Compiler stack layers

At the highest level, the compiler consists of a few key components – the graph compiler, the kernel compiler, and the kernel library.

Here’s what happens when you compile and run a model:

  1. The PyTorch model, via SambaNova Python API, goes through the graph compiler to transform the original model into a series of RDU kernel dataflow graphs.

  2. The kernel compiler is then responsible for transforming each kernel dataflow graph into a bitfile, and compiling these bitfiles together into an executable (PEF file).

  3. The user invokes the run command to perform a training or inference run, and passes in the PEF file as input.

The following illustration shows the compiler stack on the left and the runtime stack on the right.

Layers of the compiler stack

Here’s an overview of the components:

  • Samba. A Python-based software layer responsible for model graph tracing, invoking the SambaFlow compiler, and orchestrating model execution.

  • Graph compiler. Responsible for model-level graph transformation and optimizations. Transforms an ML model graph into a graph that consists of RDU kernels and execution schedules.

  • Kernel compiler. Responsible for transforming RDU kernel graphs into a PEF file. A PEF file is then used as the input to the training step.

  • Kernel library. A set of ML operator kernels (e.g. Gemm) that supports both the graph compiler and the kernel compiler. The library includes PyTorch operators that are optimized for RDU. See SambaNova PyTorch operator support for an overview.

How model compilation works

During model compilation, the SambaFlow components go through several steps, in sequence.

  • When you compile for training, the default, the process includes all steps.

  • When you compile for inference (by passing in the --inference argument), the compiler skips step 3, where the compiler computes the gradient of the loss function automatically using autograd and adds an optimizer function.

In the following explanation, each node is a PyTorch operator (Gemm, Relu, etc).


comp samba step

Samba traces the model and constructs its model graph. Each node in the diagram on the left is a PyTorch operator.


comp graph decomp

During compilation, the graph compiler performs graph decomposition.

Users can control how the compiler breaks up the graph into subgraphs with the following parameters:

  • o0 (green) supports operator-by-operator execution. o0 is the best option when you first bring up a model for debugging, but performance is not optimal.

  • o1 (red) supports operator fusion on top of operator-by-operator execution.

  • o3 (blue) is the current default. The compiler treats the entire model as one large graph. With this option, the result is faster model execution, but some models do not complete compilation.

  • o3+HD (not shown here) is manual optimization where a Human Decision (HD) file overrides certain compiler transformation and optimization decisions. HD files currently are provided only by SambaNova Customer Support.

See Arguments for compile for details.


comp training support

To support training, the compiler computes the gradient of the loss function automatically using autograd and adds an optimizer function.


comp graph optimization

Next, the graph compiler performs graph optimizations including tensor transformations and optimizations, operator parallelizations, and peephole optimizations.


comp graph mapping

During graph mapping, each function is mapped into 1 or more sections (the compiler lowers each function into 1 or more section). One section is the basic unit of compute that runs on the RDU at a time.


comp graph lowering

During graph lowering, the graph compiler lowers from mathematical operators to kernel operators. As part of this process, the compiler performs dataflow pipeline construction and optimizations.

comp dataflow optimize


comp kernel compiler

Next, the kernel compiler maps the PCU/PMU graphs onto the physical PCU/PMU instances, and generates bitfiles and the PEF file. The PEF file can then be used as the input to the first training run.

comp generate pef

comp pef components

Learn more!