Navigate:
ONNX Runtime
~$ONNXR0.5%

ONNX Runtime: ML inference and training accelerator

Cross-platform engine for optimized ONNX model execution.

LIVE RANKINGS • 10:20 AM • STEADY
OVERALL
#157
59
AI & ML
#61
17
30 DAY RANKING TREND
ovr#157
·AI#61
STARS
19.4K
FORKS
3.7K
7D STARS
+90
7D FORKS
+29
Tags:
See Repo:
Share:

Learn more about ONNX Runtime

ONNX Runtime is a runtime engine that executes machine learning models in the ONNX format across different platforms and hardware configurations. It accepts models from frameworks like PyTorch, TensorFlow, and scikit-learn, converting them to an intermediate representation for optimized execution. The runtime applies graph transformations, operator fusions, and hardware-specific optimizations to improve performance. Common deployment scenarios include inference serving in production environments and distributed training on GPU clusters.

ONNX Runtime

1

Multi-framework model support

Accepts trained models from PyTorch, TensorFlow, scikit-learn, LightGBM, XGBoost, and other libraries through the ONNX format, allowing a single runtime to handle models from different ecosystems.

2

Hardware abstraction layer

Provides consistent inference and training APIs across CPUs, GPUs, and specialized accelerators while handling platform-specific optimizations internally, reducing the need for framework-specific deployment code.

3

Graph-level optimization

Performs operator fusion, constant folding, and other graph transformations at the ONNX level before execution, independent of the original training framework used to create the model.


import onnxruntime as ort
import numpy as np

session = ort.InferenceSession("model.onnx")
input_name = session.get_inputs()[0].name
input_data = np.random.randn(1, 3, 224, 224).astype(np.float32)

outputs = session.run(None, {input_name: input_data})
print(outputs[0])


[ EXPLORE MORE ]

Related Repositories

Discover similar tools and frameworks used by developers