PyTorch Frontend for HPVM

torch2hpvm is a PyTorch frontend for HPVM. It provides a set of API that

  • Generates a PyTorch module into HPVM-C code;

  • Exports a PyTorch dataset to ApproxHPVM dataset format;

  • Compiles the generated code into binary by invoking HPVM automatically.


pip3 is the recommended package manager (also available within conda). Using pip3:

pip3 install -e ./

Getting Started

Let’s look at an example that uses DNNs and weights pre-shipped with HPVM. This is found at hpvm/benchmarks/dnn_benchmarks/pytorch/ Note that below we’ll be working under directory hpvm/benchmarks/dnn_benchmarks/pytorch.

We’ll be generating ResNet-18 into an HPVM-compiled binary. First, prepare 2 datasets for autotuning and testing.

from torch2hpvm import BinDataset
from pathlib import Path

data_dir = Path(__file__).parent / "../model_params/resnet18_cifar10"
dataset_shape = 5000, 3, 32, 32
tuneset = BinDataset(data_dir / "tune_input.bin", data_dir / "tune_labels.bin", dataset_shape)
testset = BinDataset(data_dir / "test_input.bin", data_dir / "test_labels.bin", dataset_shape)

BinDataset is a dataset created over files of ApproxHPVM dataset format. Any instance can be used here.

Note that each module is bound to 2 datasets: a “tune” and a “test” set. The generated binary accepts an argument to be either the string “tune” or “test”, and performs inference over a dataset accordingly. This is because the dataset can contain arbitrary Python code which cannot yet be exported into HPVM-C; instead the frontend has to export some predefined datasets for the model to use.

Create a DNN module and load the checkpoint:

import torch
from torch.nn import Module
import dnn  # Defined at `hpvm/benchmarks/dnn_benchmarks/pytorch`

model: Module = dnn.ResNet18()
checkpoint = Path(__file__).parent / "../model_params/resnet18_cifar10.pth.tar"

Any torch.nn.Module can be similarly used, as long as they only contain the tensor operators supported in HPVM (see “Supported Operators” and TODOs (2)).

Now we are ready to export the model. The main functioning class of torch2hpvm is ModelExporter:

from torch2hpvm import ModelExporter

output_dir = Path("./resnet18_hpvm")
build_dir = output_dir / "build"
target_binary = build_dir / "resnet18"
batch_size = 500
conf_file = "" # Change this to point to your configuration file.
exporter = ModelExporter(model, tuneset, testset, output_dir, config_file=conf_file)
exporter.generate(batch_size=batch_size).compile(target_binary, build_dir)

output_dir, build_dir, and target_binary define the folder for code generation, compilation, and path to the compiled binary respectively. batch_size is the batch size the binary uses during inference.

Note that conf_file is the path to an HPVM approximation configuration file. This file decides what approximation the binary will use during inference. This path is hardcoded into the binary and is only read when the binary starts, so it’s fine to have conf_file point to a non-existing path. An example can be found at test/dnn_benchmarks/hpvm-c/benchmarks/resnet18_cifar10/data/tuner_confs.txt.

Supported Operators

Any builtin and custom PyTorch Module are supported as long as the generated ONNX model consists of only the following operators when the Module is exported into ONNX:




















This choice of operators is largely constrained by the operators supported by the current backends (tensor_runtime).