Skip to content

when i use onnx and CUDAExecutionProvider , why onnx fp16 slow than fp32 on nvidia3090 #13628

@slantingsun

Description

@slantingsun

Search before asking

Question

python 3.8
cuda 11.3
onnxruntime-gpu 1.14.1

python export.py --weights ./best.pt --imgsz 960 --device 0 --half --simplify

Another question, will onnx using CUDAExecutionProvider be faster than pt file using gpu?
Thanks for your answer

Additional

No response

Metadata

Metadata

Assignees

No one assigned

    Labels

    exportsModel exports (ONNX, TensorRT, TFLite, etc.)questionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions