358 questions
1
vote
0
answers
52
views
OnnxRuntime with ACL Execution Provider on RK3588 (Mali-G610): Nodes assigned to ACL but GPU load remains 0%
[Goal & Problem]
I am trying to accelerate ONNX model inference on an RK3588 (Orange Pi 5) board using the Mali-G610 GPU. I have built OnnxRuntime (ORT) with the ACL (Compute Library) Execution ...
0
votes
0
answers
43
views
Fail to Run OnnxRuntime Session in C# with CUDA Device
I created a YOLO v11 DL model using a given dataset and both the train and prediction modules work fine in PyCharm.
However, after exporting the model file to ONNX format and running the prediction ...
0
votes
0
answers
36
views
onnxruntime-node native binding not found when packaging with pkg — need official way to load from custom path
I'm trying to bundle a Node.js CLI tool that uses @xenova/transformers into a single executable using pkg.
The build works fine, but when I run the packaged executable, I get this error:
Error: Cannot ...
0
votes
0
answers
94
views
Could not find a version that satisfies the requirement onnxruntime-gpu==1.22.0
currently using both python and typescript for the motia based project (https://github.com/MotiaDev/motia), python mainly for working with onnxruntime-gpu.
The OS environment is ubuntu 20.04.
firstly ...
1
vote
0
answers
84
views
Unable to load onnxruntime.dll error 0x8007045A (ERROR_DLL_INIT_FAILED)
I am having trouble using onnxruntime.dll in ML.NET.
My code works on one PC but not another. Clearly there is a mismatch between versions of some native DLLs, but I can't figure it out.
The error is:...
0
votes
1
answer
165
views
In old version (v1.18.1) of onnx web runtime, how do I specify custom location for the `.js` files (like `ort-wasm-threaded.js`)?
I have a website that runs an onnx model (pre-optimized to .ort format) on the client via ONNX's web runtime.
I must also stick to an old version of onnx (before 1.19.0), because the support for non-...
2
votes
1
answer
266
views
How to make a python package that can have two different version of a dependency?
It is now not uncommon to have a python package that is distributed in a multitude of different "flavors". This happens often with machine learning packages, e.g. onnxruntime has many "...
0
votes
0
answers
36
views
Unable to run inference in onnxruntime-web (wasm backend) proxy mode with multi-threading enabled
Has anyone had any success running ort-web on a wasm backend with the proxy option (ort.env.wasm.proxy) set and multi-threading enabled?
This is all the javascript I'm running:
// alt.ts
import * as ...
0
votes
2
answers
990
views
Running labelme in VS Code terminal gets ImportError: DLL load failed while importing onnxruntime_pybind11_state [...]
I'm following along with this YouTube video.
I've already run the command
pip install labelme opencv-python tensorflow matplotlib albumentations
in the VS Code powershell terminal and successfully ...
0
votes
0
answers
196
views
How to load onnxruntime.dll in C++ project from application directory instead of system?
I am trying to use ONNX Runtime 1.22.0 which is in my application directory. I have added Include Directories, Library Directories, Linker, and command to copy dll to application directory under Pre-...
0
votes
0
answers
88
views
Unable to get onnxruntime-web working in content.js in Chrome Extension
My file structure to show what files are produced in dist when building the project:
.
├── dist
│ ├── 97d8ed7170daf34c90d5.wasm
│ ├── content.js
│ ├── content.js.LICENSE.txt
│ ├── content.js....
1
vote
1
answer
475
views
The docling_core library (with smoldocling) fails to export to markdown
I'm trying to extract a table from a jpeg image using Smoldocling.
The result is great in 'doctags' format, but I can't export it : I don't have any warning nor error message, only an empty output.
I ...
0
votes
1
answer
3k
views
ONNXRuntime cannot load `cudnn64_9.dll` in .NET
I want to use Microsoft.ML.OnnxRuntime.Gpu version 1.20.1 on this tutorial on the ONNX website in dotnet 9.0.300. the tutorial uses CPU but I want to modify it to the GPU. However my modified code is ...
2
votes
3
answers
409
views
onnxruntime with java and spring boot cannot load onnxruntime.dll
I need to use the onnxruntime with Java and spring boot to work with mkask detection model. but I have the below exception when calling OrtEnvironment env = OrtEnvironment.getEnvironment()
I tried to ...
0
votes
0
answers
122
views
Post-processing Segmentation Mask from ONNX Model in Web (ONNX Runtime JS)
I'm currently working on running my custom ONNX segmentation model directly in the browser using ONNX Runtime for Web with HTML and JavaScript. The model is loading and running inference successfully. ...
0
votes
1
answer
87
views
Import onnxruntime then load_dataset "causes ImportError: /lib/x86_64-linux-gnu/libstdc++.so.6: version `GLIBCXX_3.4.29' not found": why + how to fix?
Running
import onnxruntime as ort
from datasets import load_dataset
yields the error:
(env-312) dernoncourt@pc:~/test$ python SE--test_importpb.py
Traceback (most recent call last):
File "/...
0
votes
0
answers
277
views
onnxruntime package missing when using silero-vad
When I use silero-vad to reduce noise for my faster-whisper model, it reports the error below:
---------------------------------------------------------------------------
ImportError ...
0
votes
0
answers
91
views
Unable to get transcriptions with the Multilingual Whisper-Small Model via Android Studio #513
Onnxruntime-inference-examples problem.
I've been encountering issues with the multilingual Whisper-small model running on ONNX Runtime in my Android application. While the app launches successfully ...
0
votes
0
answers
25
views
ONNXRuntime fails to load CUDAExecutionProvider, falls back to CPUExecutionProvider [duplicate]
I'm using InsightFace for face recognition on Windows with Python. When I initialize FaceAnalysis(), I get an error saying CUDAExecutionProvider couldn't be loaded, and it falls back to ...
0
votes
0
answers
255
views
How do I convert ONNX float32 model to bfloat16?
I have found several documentation and tools for converting ONNX model to float16, but none of them supports converting to bfloat16.
The model is originally trained using tensorflow and converted to ...
0
votes
0
answers
84
views
x.split is not a function (it is undefined) issue with Transformers js on React Native
I’m interested in using the TinyLlama-1.1B-Chat-v1.0 model with the react-native-transformers library on expo react native. The model works best with ONNX runtime and is more efficient on a Python ...
0
votes
0
answers
272
views
How to run an ONNX model on raspberry pi 4 model b running a x64 bit OS?
I have trained a DL model on matlab and exported it as an ONNX file, now I need to run it on a raspberry pi 4 model B with Raspberry Pi OS X64 bit (aarch64) and python 3.11.2, I have followed many ...
0
votes
0
answers
10
views
onnxruntime-react-native:compileDebugAidl issue
How to resolve this issue i've tried
What went wrong:
Could not determine the dependencies of task ':onnxruntime-react-native:compileDebugAidl'.
> Could not resolve all task dependencies for ...
0
votes
0
answers
44
views
previsions with a ONNX model in a C# console application with ML.NET
I'm trying to make previsions with a ONNX model in a C# console application with ML.NET
I'm getting the message: System.ArgumentException: 'Length of memory (10) must match product of dimensions (0).'
...
1
vote
1
answer
204
views
CMake '--install' installs wrong configuration
I've built ONNX Runtime for Windows from scratch by cloning https://github.com/microsoft/onnxruntime and running the build.bat script. It successfully called down to CMake and compiled and generated a ...
0
votes
0
answers
89
views
Onnxruntime quantization script for MatMulNbits, what is the type after conversion?
In the onnxruntime documentation, for quantization here:
https://onnxruntime.ai/docs/performance/model-optimizations/quantization.html#quantize-to-int4uint4
It sets accuracy_level=4 which means it's a ...
0
votes
2
answers
344
views
Yocto Bitbake Remove or Disable .debug Folder Being Created
In my bitbake yocto development environment, I'm trying to get the onnxruntime library to not add this .debug folder to our development images as its too large for our devices we're building for.
/usr/...
2
votes
0
answers
606
views
How to correctly change layers in onnx model and restore them in onnxruntime
I want to change onnx model and then restore the weights when using it with onnxruntime. However, the model doesn't seem to be changed.
First, I load the existing model and change the weights based on ...
0
votes
1
answer
77
views
Is ONNX Runtime compatible with AOT in .NET 8, and if not, is it supported in .NET 9?
I'm currently working on a .NET 8 project with the following setup:
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<OutputType>Exe</OutputType>
...
0
votes
1
answer
169
views
HuggingFace Model - OnnxRuntime - Jupyter Notebook Print Model Summary
Thank you very much for reading my question , sorry if it is an obvious question.
I use anaconda navigator : piped install the model whisper from OpenAi, which is an audio to text transformer model, I ...
0
votes
0
answers
117
views
Properly configuring use_IO_bindings in ONNX ORTModelForSequenceClassification to improve inference speed on GPU
I'm currently running into an issue where running the changes (in green) in the following diff leads to worse performance on GPU (it adds an additional 100ms/request on average, with significantly ...
0
votes
1
answer
223
views
Debugging "CUDA kernel not found in registries for Op type" in onnxruntime
My model runs much slower in onnx than in torch. During the session initialization, I get some of these messages.
[I:onnxruntime:, cuda_execution_provider.cc:2517 GetCapability] CUDA kernel not found ...
0
votes
0
answers
321
views
How do I use ONNX to detect objects using a custom yolov5s model?
Using the yolov5 ultralytics repository, I trained a yolov5s model to detect certain objects using a dataset annotated in Roboflow with one object class.
This model works great in a python repo when ...
0
votes
0
answers
386
views
Make ONNX inference faster in C#
I am using ONNX to export a model trained in Python and run it in C#.
Everything works rather well but I'd like to speed up the inference on the C# side, using Tasks or Threads on a CPU. I don't have ...
0
votes
0
answers
29
views
The onnx inference result is inconsistent with the numpy inference result
I want to implement the inference of onnx model in my own C code,but in some layers,the result between C and ONNX has 1 error, such as C is 40 but onnx is 41.
I want to know why numpy's result is -87 ...
1
vote
1
answer
74
views
Convert pythorch pth model to onnx with fixed width and variable height
My code uses PyTorch to perform segmentation annotations on PNG images. The input images have a width of 512 pixels or a multiple of this, but the height can range from 400 to 900 pixels. The code, ...
0
votes
1
answer
100
views
cmake cant find onnx runtime lib
Here is myCMakeLists.txt
cmake_minimum_required(VERSION 3.19)
project(TesteVoxarApp LANGUAGES CXX)
find_package(Qt6 6.5 REQUIRED COMPONENTS Core Widgets LinguistTools)
find_path(...
1
vote
0
answers
336
views
How to deploy llm model with golang(onnxruntime_go)?
I'm trying to deploy in production a LLM model with onnxruntime_go(https://github.com/yalue/onnxruntime_go).
The problem is that I can not get the correctly answer, since I have loaded a tokenizer and ...
0
votes
0
answers
88
views
Is it possible to override standard onnxruntime operators in C++ with runtime registration?
Referencing the onnxruntime documentation on custom operators ( https://onnxruntime.ai/docs/reference/operators/add-custom-op.html ), it's not clear to me whether:
(A) Custom operators can be ...
1
vote
2
answers
139
views
"'.../src/onnx/gen_proto.py' is not a Conan command" when building onnx
I have the following conanfile.py in my project on which I run the command conan install . --output-folder=build --build=missing -v.
import os
from conan import ConanFile
class ProjectRecipe(...
-1
votes
1
answer
112
views
Trouble building onnxruntime python bindings on macos
I am curious if anyone is doing this successfully, and if so how?
I can build the shared libraries for inference successfully using the instructions on
https://onnxruntime.ai/docs/build/inferencing....
0
votes
0
answers
763
views
ONNX Runtime CUDAExecutionProvider Fails to Load on Windows: "LoadLibrary failed with error 126" despite proper CUDA and cuDNN Installation
I'm encountering an issue when trying to run a model with ONNX Runtime using GPU acceleration on Windows. The error message indicates that the CUDAExecutionProvider cannot be loaded due to a ...
1
vote
0
answers
214
views
Run precompiled-qnn-onnx llama2 llm on Qualcomm linux
My need is to use python and load/use LLAMA as LLM. I'm right now targeting llama2-7b.
I'd like to ask which is the preferred way of executing such model on qcs 6490 using a python framework like.
I ...
0
votes
0
answers
93
views
I'm failing to run a project with a reference to onnxruntime 1.10/1.16
I'm having a C# / .NET 4.7.2 project, running onnx runtime 1.10 for a long time.
In the last year I started getting errors on some computers. I'm not sure where and why. Last week, it happens almost ...
0
votes
1
answer
751
views
Has anyone successfully converted a PyTorch-based LLM to ONNX Runtime to reduce file size?
I’m exploring the idea of converting PyTorch and TensorFlow models to ONNX Runtime to optimize performance and reduce file sizes. Since most large language models (LLMs) are built using PyTorch, I was ...
2
votes
0
answers
354
views
Why is Qualcomm's QNN x86_64 CPU backend 88x slower than onnxruntime?
I am new to Qualcomm's AI Engine Direct SDK (QNN). Before going directly to deploying AI models to Qualcomm's device, I would like to take a look at QNN's x86_64 backend, which is also relevant to QNN'...
1
vote
2
answers
507
views
Python set path
I am trying to get my gpu to work with comfyui but I keep getting errors saying
--- Applied providers: ['CPUExecutionProvider'], with options: {'CPUExecutionProvider': {}}
find model: E:\...
0
votes
1
answer
135
views
onnxruntime::qnn::QnnModel::ParseGraphInputOrOutput Dynamic shape is not supported yet, for output: /model.12/Concat_output_0_QuantizeLinear_Output
I am trying to run yolov5 onnx model in Qualcomm NPU using OnnxRuntime QNN execution provider.
I am getting this error while intializing the sesssion. ort.InferenceSession()
ParseGraphInputOrOutput ...
0
votes
1
answer
242
views
How to free webgpu gpu mem in onnxruntime web
I use onnxruntime web with following code
/**
*
* @param model don't pass session but pass model path and create session in infer inner. In this way, after infer finish, it will auto free gpu mem to ...
1
vote
1
answer
4k
views
Update on ONNXRuntimeError: LoadLibrary failed with error 126 onnxruntime\capi\onnxruntime_providers_cuda.dll
After countless tries, I got confused in this issue but finally solved it. Now, back to the initial problem that made me lose my hair: CUDA does not seem to be used when I run my model with pytorch 2....