Convert ONNX* GPT-2 Model

Public pre-trained GPT-2 model is a large transformer-based language model with a simple objective: predict the next word, given all of the previous words within some text.

Download the Pre-Trained Base GPT-2 Model

To download the model, click Download on https://github.com/onnx/models/blob/master/text/machine_comprehension/gpt-2/model/gpt2-10.onnx.

To download the model and sample test data, click Download on https://github.com/onnx/models/blob/master/text/machine_comprehension/gpt-2/model/gpt2-10.tar.gz.

Convert ONNX* GPT-2 Model to IR

To generate the Intermediate Representation (IR) of the model GPT-2, run the Model Optimizer with the following parameters:

mo --input_model gpt2-10.onnx --input_shape [X,Y,Z] --output_dir <OUTPUT_MODEL_DIR>