openvino_genai#

Classes

ContinuousBatchingPipeline

DecodedResults

EncodedResults

GenerationConfig

GenerationConfig parameters max_length: the maximum length the generated tokens can have. Corresponds to the length of the input prompt + max_new_tokens. Its effect is overridden by max_new_tokens, if also set. max_new_tokens: the maximum numbers of tokens to generate, excluding the number of tokens in the prompt. max_new_tokens has priority over max_length. ignore_eos: if set to true, then generation will not stop even if <eos> token is met. eos_token_id: token_id of <eos> (end of sentence).

GenerationResult

LLMPipeline

SchedulerConfig

StopCriteria

StopCriteria controls the stopping condition for grouped beam search. The following values are possible:

StreamerBase

TokenizedInputs

Tokenizer

openvino_genai.Tokenizer object is used to initialize Tokenizer if it's located in a different path than the main model.