import.rst 7.52 KB

Import a model

:ref:`from_onnx`

nGraph APIs can be used to run inference on a model that has been exported from a Deep Learning framework. An export produces a file with a serialized model that can be loaded and passed to one of the nGraph backends.

Importing a model from ONNX

The most-widely supported :term:`export` format available today is ONNX. Models that have been serialized to ONNX are easy to identify; they are usually named <some_model>.onnx or <some_model>.onnx.pb. These tutorials from ONNX describe how to turn trained models into an .onnx export.

Important

If you landed on this page and you already have an .onnx or .onnx.pb formatted file, you should be able to run the inference without needing to dig into anything from the "Frameworks" sections. You will, however, need to have completed the steps outlined in our :doc:`../buildlb` guide. If you intend to build nGraph for : doc:distributed-training, you will need to build that has already been compiled with the additional cmake flag: -DNGRAPH_DISTRIBUTED_ENABLE=TRUE.

To demonstrate functionality, we'll use an already-serialized CIFAR10 model trained via ResNet20. Remember that this model has already been trained and exported from a framework such as Caffe2, PyTorch or CNTK; we are simply going to build an nGraph representation of the model, execute it, and produce some outputs.

Installing ngraph_onnx with nGraph from scratch

To use ONNX models with nGraph, you will also need the companion tool ngraph_onnx, which requires Python 3.4 or higher. If nGraph has not yet been installed to your system, you can follow these steps to install everything at once; if an ngraph_dist is already installed on your system, skip ahead to the next section, :ref:`install_ngonnx`.

  1. Install prerequisites for the system and install nGraph as ngraph_dist:

    $ apt update
    $ apt install python3 python3-pip python3-dev python3-venv
    $ apt install build-essential cmake curl clang-3.9 git zlib1g zlib1g-dev libtinfo-dev
    $ git clone https://github.com/NervanaSystems/ngraph.git
    $ cd ngraph && mkdir build
    $ cd build && cmake ../ -DCMAKE_INSTALL_PREFIX=~/ngraph_dist -DNGRAPH_USE_PREBUILT_LLVM=TRUE
    $ make install
    
  2. Build the Python package (binary wheel) for ngraph and set up an env for ONNX; be sure to export the NGRAPH_CPP_BUILD_PATH where the ngraph_dist was installed.

    $ cd ngraph/python
    $ git clone --recursive -b allow-nonconstructible-holders https://github.com/jagerman/pybind11.git
    $ export PYBIND_HEADERS_PATH=$PWD/pybind11
    $ export NGRAPH_CPP_BUILD_PATH=~/ngraph_dist
    $ python3 setup.py bdist_wheel
    $ cd .. python3 -m venv onnx
    $ cd onnx/
    $ . bin/activate
    
  3. Check for the binary wheel file under /ngraph/python/dist/ and install it with pip.

    (onnx)$ pip install -U python/dist/ngraph-0.5.0-cp35-cp35m-linux_x86_64.whl
    
  4. Confirm ngraph is properly installed through a Python interpreter:

    (onnx)$ python3
    
    import ngraph as ng
    ng.abs([[1, 2, 3], [4, 5, 6]])
    <Abs: 'Abs_1' ([2, 3])>
    

    If you don't see any errors, ngraph should be installed correctly.

Installing ngraph-onnx

Add the dependencies for ONNX:

$ apt install protobuf-compiler libprotobuf-dev

Install the ngraph-onnx companion tool using pip:

(onnx) $ pip install git+https://github.com/NervanaSystems/ngraph-onnx/

Importing a serialized model

With the dependencies added, we can now import a model that has been serialized by ONNX, interact locally with the model by running Python code, create and load objects, and run inference.

This section assumes that you have your own ONNX model. With this example model from Microsoft*'s Deep Learning framework, CNTK, we can outline the procedure to show how to run ResNet on model that has been trained on the CIFAR10 data set and serialized with ONNX.

(Optional) Localize your export to the virtual environment

For this example, let's say that our serialized file was output under our $HOME directory, say at ~/onnx_conversions/trained_model.onnx. To make loading this file easier, you can run the example below from your Venv in that directory. If you invoke your python interpreter in a different directory, you will need to specify the relative path to the location of the .onnx file.

Important

If you invoke your Python interpreter in directory other than where you outputted your trained model, you will need to specify the relative path to the location of the .onnx file.

(onnx) $ cd ~/onnx_conversions
(onnx) $ python3

Enable ONNX and load an ONNX file from disk

Convert an ONNX model to an ngraph model

The importer returns a list of ngraph models for every ONNX graph output:

print(ng_models)
[{
    'name': 'Plus5475_Output_0',
    'output': <Add: 'Add_1972' ([1, 10])>,
    'inputs': [<Parameter: 'Parameter_1104' ([1, 3, 32, 32], float)>]
 }]

The output field contains the ngraph node corrsponding to the output node in the imported ONNX computational graph. The inputs list contains all input parameters for the computation which generates the output.

Using ngraph_api, create a callable computation object

Load or create an image

Run ResNet inference on picture

Put it all together

Outputs will vary greatly, depending on your model; for demonstration purposes, the code will look something like:

array([[ 1.312082 , -1.6729496,  4.2079577,  1.4012241, -3.5463796,
      2.3433776,  1.7799224, -1.6155214,  0.0777044, -4.2944093]],
   dtype=float32)