Converting an ONNX GPT-2 Model¶
Public pre-trained GPT-2 model is a large transformer-based language model with a simple objective: predict the next word, given all of the previous words within some text.
Downloading the Pre-Trained Base GPT-2 Model¶
To download the model, go to this model, and press Download.
To download the model and sample test data, go to this model, and press Download.
Converting an ONNX GPT-2 Model to IR¶
Generate the Intermediate Representation of the model GPT-2 by running model conversion with the following parameters:
mo --input_model gpt2-10.onnx --input_shape [X,Y,Z] --output_dir <OUTPUT_MODEL_DIR>