ONNXSequential: Chain Multiple ONNX Models with Ease¶
The ONNXSequential class is a powerful new feature that allows users to effortlessly combine and chain multiple ONNX models together. This is especially useful when you have several pre-trained models or custom ONNX operators that you want to execute sequentially as part of a larger pipeline.
Whether you’re working with models for inference, experimentation, or optimization, ONNXSequential makes it easier to manage, combine, and run ONNX models in a streamlined manner. It also supports flexibility in execution environments with ONNXRuntime’s execution providers (CPU, CUDA, etc.).
Key Features¶
Seamless Model Chaining: Combine multiple ONNX models into a single computational graph.
Flexible Input/Output Mapping: Control how the outputs of one model are passed as inputs to the next.
Export to ONNX: Save the combined model into a single ONNX file for easy deployment and sharing.
PyTorch-like Interface: Use the ONNXSequential class like a PyTorch nn.Sequential model, including calling it directly for inference.
Optimized Execution¶
ONNXRuntime: Automatically create optimized ONNXRuntime sessions to speed up inference.
Execution Providers Support: Utilize ONNXRuntime’s execution providers (e.g., CUDAExecutionProvider, CPUExecutionProvider, TensorrtExecutionProvider, OpenVINOExecutionProvider) for accelerated inference on different hardware.
Concurrent Sessions: You can manage multiple inference sessions concurrently, allowing for parallel processing of multiple inputs.
Asynchronous API: We offer asyncio-based execution along with the runtime’s asynchronous functions to perform non-blocking inference.
Quickstart Guide¶
Here’s how you can quickly get started with ONNXSequential:
Install ONNX and ONNXRuntime
If you haven’t already installed onnx and onnxruntime, you can install them using pip:
pip install onnx onnxruntime
Combining ONNX Models
You can initialize the ONNXSequential with a list of ONNX models or file paths. Models will be automatically chained together and optimized for inference.
import numpy as np from kornia.onnx import ONNXSequential # Initialize ONNXSequential with two models, loading from our only repo onnx_seq = ONNXSequential( "hf://operators/kornia.color.gray.RgbToGrayscale", "hf://operators/kornia.geometry.transform.affwarp.Resize_512x512" ) # Prepare some input data input_data = np.random.randn(1, 3, 256, 512).astype(np.float32) # Perform inference outputs = onnx_seq(input_data) # Print the model outputs print(outputs)
Note
By default, we assume each ONNX model contains only one input node named “input” and one output node named “output”. For complex models, you may need to pass an io_maps argument.
Input/Output Mapping Between Models
When combining models, you can specify how the outputs of one model are mapped to the inputs of the next. This allows you to chain models in custom ways.
io_map = [("model1_output_0", "model2_input_0"), ("model1_output_1", "model2_input_1")] onnx_seq = ONNXSequential("model1.onnx", "model2.onnx", io_map=io_map)
Exporting the Combined Model
You can easily export the combined model to an ONNX file:
# Export the combined model to a file onnx_seq.export("combined_model.onnx")
Optimizing with Execution Providers
Leverage ONNXRuntime’s execution providers for optimized inference. For example, to run the model on a GPU:
# Initialize with CUDA execution provider onnx_seq = ONNXSequential( "hf://operators/kornia.geometry.transform.flips.Hflip", # Or you may use a local model with either a filepath "YOUR_OWN_MODEL.onnx" or a loaded ONNX model. "hf://models/kornia.models.detection.rtdetr_r18vd_640x640", providers=['CUDAExecutionProvider'] ) # Run inference outputs = onnx_seq(input_data)
Frequently Asked Questions (FAQ)¶
1. Can I chain models from different sources?
Yes! You can chain models from different ONNX files or directly from onnx.ModelProto objects. ONNXSequential handles the integration and merging of their graphs.
2. What happens if the input/output sizes of models don’t match?
You can use the io_map parameter to control how outputs of one model are mapped to the inputs of the next. This allows for greater flexibility when chaining models with different architectures.
3. Can I use custom ONNXRuntime session options?
Absolutely! You can pass your own session options to the create_session method to fine-tune performance, memory usage, or logging.
4. How to run with CUDA?
For using CUDA ONNXRuntime, you need to install onnxruntime-gpu. For handling different CUDA version, you may refer to https://github.com/microsoft/onnxruntime/issues/21769#issuecomment-2295342211.
For example, to install onnxruntime-gpu==1.19.2 under CUDA 11.X, you may install with:
pip install onnxruntime-gpu==1.19.2 --extra-index-url https://aiinfra.pkgs.visualstudio.com/PublicPackages/_packaging/onnxruntime-cuda-11/pypi/simple/
You may then convert your sequence to CUDA, such as:
import kornia
onnx_seq = ONNXSequential(
"hf://operators/kornia.geometry.transform.flips.Hflip",
"hf://models/kornia.models.detection.rtdetr_r18vd_640x640", # Or you may use "YOUR_OWN_MODEL.onnx"
)
inp = kornia.utils.sample.get_sample_images()[0].numpy()[None]
import time
onnx_seq.as_cuda()
onnx_seq(inp) # GPU warm up
start_time = time.time()
onnx_seq(inp)
print("--- GPU %s seconds ---" % (time.time() - start_time))
onnx_seq.as_cpu()
start_time = time.time()
onnx_seq(inp)
print("--- %s seconds ---" % (time.time() - start_time))
You may get a decent improvement:
--- GPU 0.014804363250732422 seconds ---
--- CPU 0.17681646347045898 seconds ---
Why Choose ONNXSequential?¶
With the increasing adoption of ONNX for model interoperability and deployment, ONNXSequential provides a simple yet powerful interface for combining models and operators. By leveraging ONNXRuntime’s optimization and execution provider capabilities, it gives you the flexibility to: - Deploy on different hardware (CPU, GPU, TensorRT, OpenVINO, etc.). - Run complex pipelines in production environments. - Combine and experiment with models effortlessly.
Whether you’re building an advanced deep learning pipeline or simply trying to chain pre-trained models, ONNXSequential makes it easy to manage, optimize, and execute ONNX models at scale.
Get started today and streamline your ONNX workflows!
API Documentation¶
- class kornia.onnx.module.ONNXModule(op, providers=None, session_options=None, cache_dir=None, target_ir_version=None, target_opset_version=None)¶
ONNXModule to wrap an ONNX operator.
- Parameters:
arg – A variable number of ONNX models (either ONNX ModelProto objects or file paths). For Hugging Face-hosted models, use the format ‘hf://model_name’. Valid model_name can be found on https://huggingface.co/kornia/ONNX_models. Or a URL to the ONNX model.
providers (optional) – A list of execution providers for ONNXRuntime (e.g., [‘CUDAExecutionProvider’, ‘CPUExecutionProvider’]). Default:
None
session_options (optional) – Optional ONNXRuntime session options for optimizing the session. Default:
None
cache_dir (optional) – The directory where ONNX models are cached locally (only for downloading from HuggingFace). Defaults to None, which will use a default kornia.config.hub_onnx_dir directory. Default:
None
target_ir_version (optional) – The target IR version to convert to. Default:
None
target_opset_version (optional) – The target OPSET version to convert to. Default:
None
- class kornia.onnx.sequential.ONNXSequential(*args, providers=None, session_options=None, io_maps=None, cache_dir=None, auto_ir_version_conversion=False, target_ir_version=None, target_opset_version=None)¶
ONNXSequential to chain multiple ONNX operators together.
- Parameters:
*args – A variable number of ONNX models (either ONNX ModelProto objects or file paths). For Hugging Face-hosted models, use the format ‘hf://model_name’. Valid model_name can be found on https://huggingface.co/kornia/ONNX_models. Or a URL to the ONNX model.
providers (optional) – A list of execution providers for ONNXRuntime (e.g., [‘CUDAExecutionProvider’, ‘CPUExecutionProvider’]). Default:
None
session_options (optional) – Optional ONNXRuntime session options for optimizing the session. Default:
None
io_maps (optional) – An optional list of list of tuples specifying input-output mappings for combining models. If None, we assume the default input name and output name are “input” and “output” accordingly, and only one input and output node for each graph. If not None, io_maps[0] shall represent the io_map for combining the first and second ONNX models. Default:
None
cache_dir (optional) – The directory where ONNX models are cached locally (only for downloading from HuggingFace). Defaults to None, which will use a default kornia.config.hub_onnx_dir directory. Default:
None
auto_ir_version_conversion (optional) – If True, automatically convert the model’s IR version to 9, and OPSET version to 17. Other versions may be pointed to by target_ir_version and target_opset_version. Default:
False
target_ir_version (optional) – The target IR version to convert to. Default:
None
target_opset_version (optional) – The target OPSET version to convert to. Default:
None
- class kornia.onnx.utils.ONNXLoader¶
Manages ONNX models, handling local caching, downloading from Hugging Face, and loading models.
# Load a HuggingFace operator ONNXLoader.load_model("hf://operators/kornia.color.gray.GrayscaleToRgb") # doctest: +SKIP # Load a local converted/downloaded operator ONNXLoader.load_model("operators/kornia.color.gray.GrayscaleToRgb") # doctest: +SKIP
- Members: