Converting an ONNX GPT-2 Model

Public pre-trained GPT-2 model is a large transformer-based language model with a simple objective: predict the next word, given all of the previous words within some text.

Downloading the Pre-Trained Base GPT-2 Model

To download the model, go to this model, and press Download.

To download the model and sample test data, go to this model, and press Download.

Converting an ONNX GPT-2 Model to IR

Generate the Intermediate Representation of the model GPT-2 by running model conversion with the following parameters:

mo --input_model gpt2-10.onnx --input_shape [X,Y,Z] --output_dir <OUTPUT_MODEL_DIR>