openvino_genai.VLMPipeline#
- class openvino_genai.VLMPipeline#
Bases:
pybind11_objectThis class is used for generation with VLMs
- __init__(self: openvino_genai.py_openvino_genai.VLMPipeline, models_path: os.PathLike, device: str, **kwargs) None#
- device on which inference will be done
VLMPipeline class constructor. models_path (os.PathLike): Path to the folder with exported model files. device (str): Device to run the model on (e.g., CPU, GPU). Default is ‘CPU’. kwargs: Device properties
Methods
__delattr__(name, /)Implement delattr(self, name).
__dir__()Default dir() implementation.
__eq__(value, /)Return self==value.
__format__(format_spec, /)Default object formatter.
__ge__(value, /)Return self>=value.
__getattribute__(name, /)Return getattr(self, name).
__gt__(value, /)Return self>value.
__hash__()Return hash(self).
__init__(self, models_path, device, **kwargs)device on which inference will be done
This method is called when a class is subclassed.
__le__(value, /)Return self<=value.
__lt__(value, /)Return self<value.
__ne__(value, /)Return self!=value.
__new__(**kwargs)Helper for pickle.
__reduce_ex__(protocol, /)Helper for pickle.
__repr__()Return repr(self).
__setattr__(name, value, /)Implement setattr(self, name, value).
Size of object in memory, in bytes.
__str__()Return str(self).
Abstract classes can override this to customize issubclass().
finish_chat(self)generate(*args, **kwargs)Overloaded function.
get_generation_config(self)get_tokenizer(self)set_chat_template(self, new_template)set_generation_config(self, new_config)start_chat(self[, system_message])- __class__#
alias of
pybind11_type
- __delattr__(name, /)#
Implement delattr(self, name).
- __dir__()#
Default dir() implementation.
- __eq__(value, /)#
Return self==value.
- __format__(format_spec, /)#
Default object formatter.
- __ge__(value, /)#
Return self>=value.
- __getattribute__(name, /)#
Return getattr(self, name).
- __gt__(value, /)#
Return self>value.
- __hash__()#
Return hash(self).
- __init__(self: openvino_genai.py_openvino_genai.VLMPipeline, models_path: os.PathLike, device: str, **kwargs) None#
- device on which inference will be done
VLMPipeline class constructor. models_path (os.PathLike): Path to the folder with exported model files. device (str): Device to run the model on (e.g., CPU, GPU). Default is ‘CPU’. kwargs: Device properties
- __init_subclass__()#
This method is called when a class is subclassed.
The default implementation does nothing. It may be overridden to extend subclasses.
- __le__(value, /)#
Return self<=value.
- __lt__(value, /)#
Return self<value.
- __ne__(value, /)#
Return self!=value.
- __new__(**kwargs)#
- __reduce__()#
Helper for pickle.
- __reduce_ex__(protocol, /)#
Helper for pickle.
- __repr__()#
Return repr(self).
- __setattr__(name, value, /)#
Implement setattr(self, name, value).
- __sizeof__()#
Size of object in memory, in bytes.
- __str__()#
Return str(self).
- __subclasshook__()#
Abstract classes can override this to customize issubclass().
This is invoked early on by abc.ABCMeta.__subclasscheck__(). It should return True, False or NotImplemented. If it returns NotImplemented, the normal algorithm is used. Otherwise, it overrides the normal algorithm (and the outcome is cached).
- finish_chat(self: openvino_genai.py_openvino_genai.VLMPipeline) None#
- generate(*args, **kwargs)#
Overloaded function.
generate(self: openvino_genai.py_openvino_genai.VLMPipeline, prompt: str, images: list[openvino._pyopenvino.Tensor], generation_config: openvino_genai.py_openvino_genai.GenerationConfig, streamer: Union[Callable[[str], bool], openvino_genai.py_openvino_genai.StreamerBase, None] = None, **kwargs) -> Union[openvino_genai.py_openvino_genai.DecodedResults]
Generates sequences for VLMs.
- param prompt:
input prompt
- type prompt:
str
- param images:
image or list of images
- type images:
List[ov.Tensor] or ov.Tensor
- param generation_config:
generation_config
- type generation_config:
GenerationConfig or a Dict
- param streamer:
streamer either as a lambda with a boolean returning flag whether generation should be stopped
:type : Callable[[str], bool], ov.genai.StreamerBase
- param kwargs:
arbitrary keyword arguments with keys corresponding to GenerationConfig fields.
:type : Dict
- return:
return results in decoded form
- rtype:
DecodedResults
generate(self: openvino_genai.py_openvino_genai.VLMPipeline, prompt: str, images: openvino._pyopenvino.Tensor, generation_config: openvino_genai.py_openvino_genai.GenerationConfig, streamer: Union[Callable[[str], bool], openvino_genai.py_openvino_genai.StreamerBase, None] = None, **kwargs) -> Union[openvino_genai.py_openvino_genai.DecodedResults]
Generates sequences for VLMs.
- param prompt:
input prompt
- type prompt:
str
- param images:
image or list of images
- type images:
List[ov.Tensor] or ov.Tensor
- param generation_config:
generation_config
- type generation_config:
GenerationConfig or a Dict
- param streamer:
streamer either as a lambda with a boolean returning flag whether generation should be stopped
:type : Callable[[str], bool], ov.genai.StreamerBase
- param kwargs:
arbitrary keyword arguments with keys corresponding to GenerationConfig fields.
:type : Dict
- return:
return results in decoded form
- rtype:
DecodedResults
generate(self: openvino_genai.py_openvino_genai.VLMPipeline, prompt: str, **kwargs) -> Union[openvino_genai.py_openvino_genai.DecodedResults]
Generates sequences for VLMs.
- param prompt:
input prompt
- type prompt:
str
- param kwargs:
arbitrary keyword arguments with keys corresponding to generate params.
Expected parameters list: image: ov.Tensor - input image, images: List[ov.Tensor] - input images, generation_config: GenerationConfig, streamer: Callable[[str], bool], ov.genai.StreamerBase - streamer either as a lambda with a boolean returning flag whether generation should be stopped
- return:
return results in decoded form
- rtype:
DecodedResults
- get_generation_config(self: openvino_genai.py_openvino_genai.VLMPipeline) openvino_genai.py_openvino_genai.GenerationConfig#
- get_tokenizer(self: openvino_genai.py_openvino_genai.VLMPipeline) openvino_genai.py_openvino_genai.Tokenizer#
- set_chat_template(self: openvino_genai.py_openvino_genai.VLMPipeline, new_template: str) None#
- set_generation_config(self: openvino_genai.py_openvino_genai.VLMPipeline, new_config: openvino_genai.py_openvino_genai.GenerationConfig) None#
- start_chat(self: openvino_genai.py_openvino_genai.VLMPipeline, system_message: str = '') None#