Error while converting the yolov8n.onnx model

Which system do you use? Android, Ubuntu, OOWOW or others?

Others

Which version of system do you use? Please provide the version of the system here:

Target: Yocto.
Host: Ubuntu 24.04

Please describe your issue below:

I took a pre-trained model (yolov8n.pt) and converted to onnx using the script + patch (export.py) given in vim3_demo_lite website. I installed docker too.

Then when I run the bash convert-in-docker.sh I get error from acuitylib/onnx_ir/onnx_numpy_backend/smart_toolkit.py file, where I search for code, then I couldn’t find it in the cloned sdk repo. But I assume it to be running inside the docker. The detailed log is listed below.

On an initial look, it looks like a compatibility issue between the version of ultralytics that runs/installed on the host vs the docker. But I didn’t look into the code in detail, and thought if can ask here to see if there are any quick fixes / solution.

Post a console log of your issue below:


(myenv) emb-aanahcn@cariqserver:~/labs/yolo/aml_npu_sdk$ bash convert-in-docker.sh
docker run -it --name npu-vim3 --rm -v /home/emb-aanahcn/labs/yolo/aml_npu_sdk:/home/khadas/npu -v /etc/localtime:/etc/localtime:ro -v /etc/timezone:/etc/timezone:ro -v /home/emb-aanahcn:/home/emb-aanahcn numbqq/npu-vim3
2024-10-14 14:42:47.525038: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/khadas/npu/acuity-toolkit/bin/acuitylib
2024-10-14 14:42:47.525071: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
I Namespace(import='onnx', input_dtype_list=None, input_size_list=None, inputs=None, model='./model/yolov8n.onnx', output_data='yolov8n.data', output_model='yolov8n.json', outputs=None, size_with_batch=None, which='import')
I Start importing onnx...
WARNING: ONNX Optimizer has been moved to https://github.com/onnx/optimizer.
All further enhancements and fixes to optimizers will be done in this new repo.
The optimizer code in onnx/onnx repo will be removed in 1.9 release.

W Call onnx.optimizer.optimize fail, skip optimize
I Current ONNX Model use ir_version 10 opset_version 19
I Call acuity onnx optimize 'eliminate_option_const' success
/home/khadas/npu/acuity-toolkit/bin/acuitylib/acuitylib/onnx_ir/onnx_numpy_backend/ops/split.py:15: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison
  if inputs[1] == '':
W Call acuity onnx optimize 'froze_const_branch' fail, skip this optimize
I Call acuity onnx optimize 'froze_if' success
I Call acuity onnx optimize 'merge_sequence_construct_concat_from_sequence' success
I Call acuity onnx optimize 'merge_lrn_lowlevel_implement' success
Traceback (most recent call last):
  File "pegasus.py", line 131, in <module>
  File "pegasus.py", line 112, in main
  File "acuitylib/app/importer/commands.py", line 245, in execute
  File "acuitylib/vsi_nn.py", line 171, in load_onnx
  File "acuitylib/app/importer/import_onnx.py", line 123, in run
  File "acuitylib/converter/onnx/convert_onnx.py", line 61, in __init__
  File "acuitylib/converter/onnx/convert_onnx.py", line 761, in _shape_inference
  File "acuitylib/onnx_ir/onnx_numpy_backend/shape_inference.py", line 65, in infer_shape
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_graph_engine.py", line 70, in smart_onnx_scanner
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_node.py", line 48, in calc_and_assign_smart_info
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_toolkit.py", line 636, in multi_direction_broadcast_shape

ValueError: operands could not be broadcast together with shapes (1,0,160,160) (1,16,160,160)


[9] Failed to execute script pegasus
2024-10-14 14:42:50.712274: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/khadas/npu/acuity-toolkit/bin/acuitylib
2024-10-14 14:42:50.712351: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
I Namespace(channel_mean_value='0 0 0 0.0039215', generate='inputmeta', input_meta_output='yolov8n_inputmeta.yml', model='yolov8n.json', separated_database=False, source_file='dataset.txt', which='generate')
I Load model in yolov8n.json
Traceback (most recent call last):
  File "pegasus.py", line 131, in <module>
  File "pegasus.py", line 123, in main
  File "acuitylib/app/console/commands.py", line 58, in execute
  File "acuitylib/acuitynet.py", line 147, in load_file
  File "acuitylib/acuitynet.py", line 510, in load
FileNotFoundError: [Errno 2] No such file or directory: 'yolov8n.json'
[76] Failed to execute script pegasus
0_import_model.sh: line 81: warning: here-document at line 24 delimited by end-of-file (wanted `COMMENT')
2024-10-14 14:42:52.793470: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/khadas/npu/acuity-toolkit/bin/acuitylib
2024-10-14 14:42:52.793510: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
I Namespace(MLE=False, algorithm='normal', batch_size=0, compute_entropy=False, device=None, divergence_first_quantize_bits=11, divergence_nbins=0, hybrid=False, iterations=1, model='yolov8n.json', model_data='yolov8n.data', model_quantize=None, moving_average_weight=0.01, output_dir=None, qtype='int8', quantizer='dynamic_fixed_point', rebuild=True, rebuild_all=False, which='quantize', with_input_meta='yolov8n_inputmeta.yml')
2024-10-14 14:42:54.173359: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations:  AVX2 FMA
To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-10-14 14:42:54.180799: I tensorflow/core/platform/profile_utils/cpu_utils.cc:104] CPU Frequency: 2245780000 Hz
2024-10-14 14:42:54.182223: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x5f0dbe0 initialized for platform Host (this does not guarantee that XLA will be used). Devices:
2024-10-14 14:42:54.182253: I tensorflow/compiler/xla/service/service.cc:176]   StreamExecutor device (0): Host, Default Version
2024-10-14 14:42:54.183577: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/khadas/npu/acuity-toolkit/bin/acuitylib
2024-10-14 14:42:54.183596: W tensorflow/stream_executor/cuda/cuda_driver.cc:312] failed call to cuInit: UNKNOWN ERROR (303)
2024-10-14 14:42:54.183618: I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:156] kernel driver does not appear to be running on this host (0c80a6be6d85): /proc/driver/nvidia/version does not exist
I Load model in yolov8n.json
Traceback (most recent call last):
  File "pegasus.py", line 131, in <module>
  File "pegasus.py", line 108, in main
  File "acuitylib/app/medusa/commands.py", line 204, in execute
  File "acuitylib/vsi_nn.py", line 280, in load_model
  File "acuitylib/acuitynet.py", line 510, in load
FileNotFoundError: [Errno 2] No such file or directory: 'yolov8n.json'
[113] Failed to execute script pegasus

Hello @aananthcn ,

Now this problem solution is downgrade the version of ultralytics and PyTorch to this version.

Then use this version to train a new model.

Could you provide your pre-trained model(yolov8n.pt) and your ultralytics, PyTorch verison that help us find the reason of problem.

1 Like

Sure, will find out how to use the specific version of the above mentioned modules for the conversion of the model.

Regarding the files and version: (note: the link is valid for 21 days only)

  • yolo8n.pt
  • Ultralytics version: 8.3.12
  • PyTorch version: 2.4.1

@Louis-Cheng-Liu, I am able to downgrade ultralytics to 8.0.86. But I can’t downgrade pytorch to 1.10.1. Here is the error log:

(myenv) emb-aanahcn@cariqserver:~/labs/yolo$ pip install torch==1.10.1
ERROR: Could not find a version that satisfies the requirement torch==1.10.1 (from versions: 1.11.0, 1.12.0, 1.12.1, 1.13.0, 1.13.1, 2.0.0, 2.0.1, 2.1.0, 2.1.1, 2.1.2, 2.2.0, 2.2.1, 2.2.2, 2.3.0, 2.3.1, 2.4.0, 2.4.1)
ERROR: No matching distribution found for torch==1.10.1

Is version 1.10.1 a right pytorch version? Should I try 1.11.0?

Hello @aananthcn ,

It is incompatible between the version of Python and PyTorch. So pip cannot find this version.

Downgrade Python version. But i suggest you trying 1.11.0 first. The modification between them are minor.

Hello @Louis-Cheng-Liu,

This is just a intermediate update. I will start to train a model from scratch as you suggested earlier.

Update:

PyTorch version 1.11.0 gave me the same ValueError as mentioned in log that I posted on 14th Oct 2024.

Also note that, for ultralytics version 8.0.86, the file myenv/lib/python3.9/site-packages/ultralytics/nn/modules/head.py doesn’t exist as mentioned in the vim3_demo_lite page. Instead I found similar lines in myenv/lib/python3.9/site-packages/ultralytics/nn/modules.py file and I did the patch.

Here are my detailed steps:

	• Host: Ubuntu 22.04.5 LTS
	• sudo add-apt-repository ppa:deadsnakes/ppa
	• sudo apt update
	• sudo apt install python3.9 python3.9-venv python3.9-dev
	• python3.9 -m venv myenv
	• source myenv/bin/activate
	• pip install torch==1.10.1 ultralytics==8.0.86
	• pip install 'numpy<2'

After the above step, when I used the old yolov8n.onnx (converted using a higher version of ultralytics), then again I got the same ValueError when I ran the convert-in-docker.sh script.

Then I decided to delete the .onnx file and tried to export the yolov8n.pt file using the export.py file. Now the export itself is failing. Detailed log output is below:

(myenv) emb-aanahcn@cariqserver:~/labs/yolo$ python export-yolo-2-onnx.py
WARNING ⚠️ ./yolov8n.pt appears to require 'ultralytics.nn.modules.conv', which is not in ultralytics requirements.
AutoInstall will run now for 'ultralytics.nn.modules.conv' but this feature will be removed in the future.
Recommend fixes are to train a new model using the latest 'ultralytics' package or to run a command with an official YOLOv8 model, i.e. 'yolo predict model=yolov8n.pt'
requirements: YOLOv8 requirement "ultralytics.nn.modules.conv" not found, attempting AutoUpdate...
ERROR: Could not find a version that satisfies the requirement ultralytics.nn.modules.conv (from versions: none)
ERROR: No matching distribution found for ultralytics.nn.modules.conv

[notice] A new release of pip is available: 23.0.1 -> 24.2
[notice] To update, run: pip install --upgrade pip
requirements: ❌ Command 'pip install --no-cache "ultralytics.nn.modules.conv"  ' returned non-zero exit status 1.
Traceback (most recent call last):
  File "/home/emb-aanahcn/labs/yolo/export-yolo-2-onnx.py", line 4, in <module>
    model = YOLO("./yolov8n.pt")
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/ultralytics/yolo/engine/model.py", line 107, in __init__
    self._load(model, task)
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/ultralytics/yolo/engine/model.py", line 156, in _load
    self.model, self.ckpt = attempt_load_one_weight(weights)
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/ultralytics/nn/tasks.py", line 417, in attempt_load_one_weight
    ckpt, weight = torch_safe_load(weight)  # load ckpt
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/ultralytics/nn/tasks.py", line 372, in torch_safe_load
    return torch.load(file, map_location='cpu'), file  # load
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/torch/serialization.py", line 607, in load
    return _load(opened_zipfile, map_location, pickle_module, **pickle_load_args)
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/torch/serialization.py", line 882, in _load
    result = unpickler.load()
  File "/home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages/torch/serialization.py", line 875, in find_class
    return super().find_class(mod_name, name)
ModuleNotFoundError: No module named 'ultralytics.nn.modules.conv'; 'ultralytics.nn.modules' is not a package
(myenv) emb-aanahcn@cariqserver:~/labs/yolo$ pip show ultralytics
Name: ultralytics
Version: 8.0.86
Summary: Ultralytics YOLOv8
Home-page: https://github.com/ultralytics/ultralytics
Author: Ultralytics
Author-email: hello@ultralytics.com
License: AGPL-3.0
Location: /home/emb-aanahcn/labs/yolo/myenv/lib/python3.9/site-packages
Requires: matplotlib, numpy, opencv-python, pandas, Pillow, psutil, PyYAML, requests, scipy, seaborn, sentry-sdk, thop, torch, torchvision, tqdm
Required-by:

Kindly help me know if any of my steps are incorrect?

Hello @aananthcn ,

After downgrading, you need to train a new model. The pretrained model is trained by new version. Some layers’ name are changed.

Hi @Louis-Cheng-Liu,

Even after creating a model and train from scratch, I get the same error. Links to full log, the new model are given in the end.

(myenv) emb-aanahcn@cariqserver:~/labs/yolo/aml_npu_sdk$ ./convert-in-docker.sh
docker run -it --name npu-vim3 --rm -v /home/emb-aanahcn/labs/yolo/aml_npu_sdk:/home/khadas/npu -v /etc/localtime:/etc/localtime:ro -v /etc/timezone:/etc/timezone:ro -v /home/emb-aanahcn:/home/emb-aanahcn numbqq/npu-vim3
2024-10-18 14:10:51.233517: W tensorflow/stream_executor/platform/default/dso_loader.cc:59] Could not load dynamic library 'libcudart.so.10.1'; dlerror: libcudart.so.10.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/khadas/npu/acuity-toolkit/bin/acuitylib
2024-10-18 14:10:51.233557: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
I Namespace(import='onnx', input_dtype_list=None, input_size_list=None, inputs=None, model='./model/yolov8n.onnx', output_data='yolov8n.data', output_model='yolov8n.json', outputs=None, size_with_batch=None, which='import')
I Start importing onnx...
WARNING: ONNX Optimizer has been moved to https://github.com/onnx/optimizer.
All further enhancements and fixes to optimizers will be done in this new repo.
The optimizer code in onnx/onnx repo will be removed in 1.9 release.

W Call onnx.optimizer.optimize fail, skip optimize
I Current ONNX Model use ir_version 7 opset_version 13
I Call acuity onnx optimize 'eliminate_option_const' success
/home/khadas/npu/acuity-toolkit/bin/acuitylib/acuitylib/onnx_ir/onnx_numpy_backend/ops/split.py:15: FutureWarning: elementwise comparison failed; returning scalar instead, but in the future will perform elementwise comparison
  if inputs[1] == '':
W Call acuity onnx optimize 'froze_const_branch' fail, skip this optimize
I Call acuity onnx optimize 'froze_if' success
I Call acuity onnx optimize 'merge_sequence_construct_concat_from_sequence' success
I Call acuity onnx optimize 'merge_lrn_lowlevel_implement' success
Traceback (most recent call last):
  File "pegasus.py", line 131, in <module>
  File "pegasus.py", line 112, in main
  File "acuitylib/app/importer/commands.py", line 245, in execute
  File "acuitylib/vsi_nn.py", line 171, in load_onnx
  File "acuitylib/app/importer/import_onnx.py", line 123, in run
  File "acuitylib/converter/onnx/convert_onnx.py", line 61, in __init__
  File "acuitylib/converter/onnx/convert_onnx.py", line 761, in _shape_inference
  File "acuitylib/onnx_ir/onnx_numpy_backend/shape_inference.py", line 65, in infer_shape
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_graph_engine.py", line 70, in smart_onnx_scanner
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_node.py", line 48, in calc_and_assign_smart_info
  File "acuitylib/onnx_ir/onnx_numpy_backend/smart_toolkit.py", line 636, in multi_direction_broadcast_shape
ValueError: operands could not be broadcast together with shapes (1,0,160,160) (1,16,160,160)

Links to important files (links will expire after 25 days):

Kindly help me find out my mistake.

Hello @aananthcn ,

I try many times that second and third link cannot download.
image

Apologies. I have updated the last 2 links in my above message. I am also providing the same links below:

In the mean time I realized that, while starting the training from scratch, the model yaml file were not optimized for int8 (I just randomly chose configs) here is that yaml file:

Hello @aananthcn ,

The ONNX link is still downloading failed.

I use your pt model to convert ONNX model and then convert nb model. It succeeds without mistake.

Here is my convert model use your pt. You can compare it with your model.
https://dl.khadas.com/.test/yolov8n-new.onnx
https://dl.khadas.com/.test/yolov8n-new.nb

And i do not find the yaml file in ultralytics codes. Is the file in ultralytics? Or you use other YOLOv8 training codes? You do not need to do quantification when train. The convert tool will quantited model according you set.

1 Like

Thank you @Louis-Cheng-Liu for support. I am new to AI so, I will learn how to compare .onnx files and then let you know if there are any differences (I strongly see that there are differences, read below for details).

But I became curious and installed your .nb file into the VIM3 board that is running my Yocto based Linux system, but I got errors related to detect apis (see the log at the bottom of this message).

But I have two questions:

  1. First question: I am getting the same error consistently with all my .onnx file. But when I use your’s .onnx file, then I see some progress. From the log you will see that I am getting a permission error now:
I End importing onnx...
I Dump net to yolov8n.json
Traceback (most recent call last):
  File "pegasus.py", line 131, in <module>
  File "pegasus.py", line 112, in main
  File "acuitylib/app/importer/commands.py", line 246, in execute
  File "acuitylib/vsi_nn.py", line 726, in save_model
  File "acuitylib/acuitynet.py", line 595, in dump
PermissionError: [Errno 13] Permission denied: 'yolov8n.json'

This indicates my setup is wrong and even my conversion to .onnx file (by export.py) itself is wrong. I will investigate it. Do you suggest any specific areas to look at (to resolve my conversion issues)


  1. Second question is on the error I get when I run the yolov8n_demo_x11_usb demo. What are the possible reason why I get these errors?
root@khadas-vim3:~# yolov8n_demo_x11_usb -d /dev/video1
W Detect_api:[det_set_log_level:19]Set log level=1
W Detect_api:[det_set_log_level:21]output_format not support Imperfect, default to DET_LOG_TERMINAL
W Detect_api:[det_set_log_level:26]Not exist VSI_NN_LOG_LEVEL, Setenv set_vsi_log_error_level
det_set_log_config Debug
E [model_create:63]CHECK STATUS(-4:The given graph has failed verification due to an insufficient number of required parameters, which cannot be automatically created. Typically this indicates required atomic parameters.)
E Detect_api:[det_set_model:226]Model_create fail, file_path=nn_data, dev_type=1
det_set_model fail. ret=-4

Hello @aananthcn ,

Aout first question, it looks like the permission question.


The file yolov8n.json is generated in docker. You can try command as follows.

sudo bash convert-in-docker.sh

About your second question, your command forget add model path.

About how to compare model. Open your ONNX model in this website Netron.


On the left is the model structure and on the right is a layer information. Click the model input(red box) and then you will see all version of this model in right(blue box).

1 Like

Thank you @Louis-Cheng-Liu.

Here is my Model property.

I see few differences. Please share your recommendations. In the mean time I will downgrade pytorch to 1.8 and update my results.

Regarding the error while running the demo app, even after passing the model file as argument (below), I get the same VX_ERROR_NOT_SUFFICIENT error when vxVerifyGraph() is invoked.

./yolov8n_demo_x11_usb -m nn_data/yolo_face_88.nb -d /dev/video1 
...
E [model_create:63]CHECK STATUS(-4:The given graph has failed verification due to an insufficient number of required parameters, which cannot be automatically created. Typically this indicates required atomic parameters.)
E Detect_api:[det_set_model:226]Model_create fail, file_path=nn_data, dev_type=1
det_set_model fail. ret=-4
...

Hello @Louis-Cheng-Liu !

Update: I am able to successfully train, convert and generate yolov8n.nb file. Here are the key steps:

  1. Installed python version 3.9
    • sudo apt install python3.9 python3.9-venv python3.9-dev
  2. Installed the following versions of python module
    • pip install torch==1.8 ultralytics==8.0.86 'numpy<2'
    • note: PyTorch == 1.10.1 is a misleading statement, 1.10.1 version of torch doesn’t work.
  3. Then follow the instructions in vim3_demo_lite/yolov8n page.
  4. Finally, modify the following line in “aml_npu_sdk/convert-in-docker.sh” file
    • bash 0_import_model.sh && bash 1_quantize_model.sh && bash 2_export_case_code.sh
      :point_down:
    • sudo bash 0_import_model.sh && sudo bash 1_quantize_model.sh && sudo bash 2_export_case_code.sh
  5. Then run the conver-in-docker.sh script to get the conversion done without error.

After this, if I run the demo application using the trained and converted file on the VIM3 board running Yocto Linux, I get the same error (below).

root@khadas-vim3:~# sudo yolov8n_demo_x11_usb -m /home/root/nn_data/yolo_face_88.nb -d /dev/video1
W Detect_api:[det_set_log_level:19]Set log level=1
W Detect_api:[det_set_log_level:21]output_format not support Imperfect, default to DET_LOG_TERMINAL
W Detect_api:[det_set_log_level:26]Not exist VSI_NN_LOG_LEVEL, Setenv set_vsi_log_error_level
det_set_log_config Debug
E [model_create:63]CHECK STATUS(-4:The given graph has failed verification due to an insufficient number of required parameters, which cannot be automatically created. Typically this indicates required atomic parameters.)
E Detect_api:[det_set_model:226]Model_create fail, file_path=nn_data, dev_type=1
det_set_model fail. ret=-4

It is good that my original question or issue is resolved. Should I raise a separate topic to discuss the run time errors?

Hello @aananthcn ,

Could you provide the log after running the commands as follows.

$ export VSI_NN_LOG_LEVEL=4
$ sudo yolov8n_demo_x11_usb -m /home/root/nn_data/yolo_face_88.nb -d /dev/video1

Here is the output:

root@khadas-vim3:~# export VSI_NN_LOG_LEVEL=4
root@khadas-vim3:~# sudo yolov8n_demo_x11_usb -m /home/root/nn_data/yolo_face_88.nb -d /dev/video1
W Detect_api:[det_set_log_level:19]Set log level=1
W Detect_api:[det_set_log_level:21]output_format not support Imperfect, default to DET_LOG_TERMINAL
W Detect_api:[det_set_log_level:26]Not exist VSI_NN_LOG_LEVEL, Setenv set_vsi_log_error_level
det_set_log_config Debug
E [model_create:63]CHECK STATUS(-4:The given graph has failed verification due to an insufficient number of required parameters, which cannot be automatically created. Typically this indicates required atomic parameters.)
E Detect_api:[det_set_model:226]Model_create fail, file_path=nn_data, dev_type=1
det_set_model fail. ret=-4
[ WARN:0@1.181] global cap_gstreamer.cpp:1777 open OpenCV | GStreamer warning: Cannot query video position: status=0, value=-1, duration=-1
[ WARN:0@1.210] global cap_gstreamer.cpp:2838 handleMessage OpenCV | GStreamer warning: Embedded video playback halted; module v4l2src0 reported: Internal data stream error.
[ WARN:0@1.211] global cap_gstreamer.cpp:1208 startPipeline OpenCV | GStreamer warning: unable to start pipeline
[ WARN:0@1.211] global cap_gstreamer.cpp:1971 setProperty OpenCV | GStreamer warning: no pipeline
capture device failed to open!
[ WARN:0@1.211] global cap_gstreamer.cpp:1173 isPipelinePlaying OpenCV | GStreamer warning: GStreamer: pipeline have not been created

Hello @aananthcn ,

Could you provide your nb model. From the log, i cannot make sure what is wrong with it.

Please find my files (this time I am sharing through different channel):

Similarly, can you share me a working or tested .nb file, it will help me to crossverify the drivers and libs on my Yocto Linux.

Hello @aananthcn ,

I have got your model. And i forget asking which kernel do you use?