Converting an ONNX GPT-2 Model


The code described here has been deprecated! Do not use it to avoid working with a legacy solution. It will be kept for some time to ensure backwards compatibility, but you should not use it in contemporary applications.

This guide describes a deprecated conversion method. The guide on the new and recommended method can be found in the Python tutorials.

Public pre-trained GPT-2 model is a large transformer-based language model with a simple objective: predict the next word, given all of the previous words within some text.

Downloading the Pre-Trained Base GPT-2 Model

To download the model, go to this model, and press Download.

To download the model and sample test data, go to this model, and press Download.

Converting an ONNX GPT-2 Model to IR

Generate the Intermediate Representation of the model GPT-2 by running model conversion with the following parameters:

mo --input_model gpt2-10.onnx --input_shape [X,Y,Z] --output_dir <OUTPUT_MODEL_DIR>