Converting an ONNX GPT-2 Model

Public pretrained GPT-2 model is a large transformer-based language model with a simple objective: predict the next word, given all of the previous words within some text.

Downloading the Pre-Trained Base GPT-2 Model

To download the model, go to this model, and press Download.

To download the model and sample test data, go to this model, and press Download.

Converting an ONNX GPT-2 Model to IR

Generate the Intermediate Representation of the model GPT-2 by running Model Optimizer with the following parameters:

mo --input_model gpt2-10.onnx --input_shape [X,Y,Z] --output_dir <OUTPUT_MODEL_DIR>