Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR adds runtimes other than the ONNX Runtime.
There are plenty of alternative ONNX inference engines for Rust that each provide their own unique qualities:
candle
boasts impressive performance and GPU acceleration.tract
is a battle-tested pure-Rust inference engine with excellent operator support.wonnx
focuses on broader GPU support is designed for the web.With the removal of
wasm32-unknown-unknown
support from ort, and the Sisyphean task of getting the damn thing to link, it's clear that ONNX Runtime isn't always the best choice. Most often, though, it is the best choice for one platform, but not another. Applications wishing to target CUDA on desktop and WebGPU on web would need to have 2 different code paths usingort
andwonnx
. Adding support forwonnx
& others directly inort
would mean that developers only need to use theort
API to target both backends, and only a single line of code is required to switch between them, e.g.ort::set_api(ort_candle::api());
Status
ort-candle
Alternative 'execution providers'- appears to currently be unsupported bycandle-onnx
?IoBinding
ort-tract
ort-wonnx
p.s., sponsorships allow me to spend more time on this PR =)