CPUs - ZenDNN

AMD Inference Server is integrated with ZenDNN optimized libraries for running inference on AMD EPYC processors with TensorFlow and PyTorch.

Build an image

To build an image with TensorFlow/Pytorch + ZenDNN enabled, you need to download the ZenDNN libraries and then build the image by pointing the build script to the location of these downloaded packages.

Download the ZenDNN packages

You can download the TF+ZenDNN and/or PT+ZenDNN packages from the ZenDNN developer downloads. Before downloading these packages, you will be required to read and agree to the End User License Agreement (EULA).

  1. For TensorFlow: download TF_v2.10_ZenDNN_v4.0_C++_API.zip

  2. For PyTorch: download PT_v1.12_ZenDNN_v4.0_C++_API.zip

You can download one or both, depending on what you want to enable.

Build the image

After downloading the packages, place them in the root of the repository. To build an image with ZenDNN enabled, you need to add the the appropriate flag(s) to the amdinfer dockerize command:

# create the Dockerfile
python3 docker/generate.py

# build the dev image $(whoami)/amdinfer-dev-zendnn:latest
./amdinfer dockerize --tfzendnn=./TF_v2.10_ZenDNN_v4.0_C++_API.zip --ptzendnn=./PT_v1.12_ZenDNN_v4.0_C++_API.zip --suffix="-zendnn"

# build the deployment image $(whoami)/amdinfer-zendnn:latest
./amdinfer dockerize --tfzendnn=./TF_v2.10_ZenDNN_v4.0_C++_API.zip --ptzendnn=./PT_v1.12_ZenDNN_v4.0_C++_API.zip --suffix="-zendnn" --production

You can choose to build a image with just TF+ZenDNN or just PT+ZenDNN by only passing the appropriate flag.

Note

The downloaded ZenDNN packages will be used by the Docker build process so they must be in the inference server repository directory and in a location that is not excluded by the .dockerignore file. These instructions suggest using the repository root but any path that meets this criteria will work.

Get assets and models

You can download the assets and models used for tests and examples with:

$ ./amdinfer get --tfzendnn --ptzendnn

Freezing PyTorch models

To use PyTorch models with the AMD Inference Server, you need to convert downloaded PyTorch Eager models to TorchScript.

Run Tests

To verify the working of TF/PT + ZenDNN in the AMD Inference Server, run a sample test case. This test will load a model and run with a sample image and assert the output.

  1. For TensorFlow + ZenDNN

    $ ./amdinfer test -k tfzendnn
    
  2. For PyTorch + ZenDNN

    $ ./amdinfer test -k ptzendnn
    

Tune performance

For tuning ZenDNN performance, you can refer to the TensorFlow + ZenDNN and PyTorch + ZenDNN user guides.