Comments (7)
Cause we donot have 2080Ti right now, we cannot reproduce your error locally. There are two ways to skip such error I used.
You can firstly test if it still get error by using default kernel.
./pplnn-build/tools/pplnn --use-cuda --onnx-model=mobilenet_v2.onnx --kernel-type=float16 --quick-select
Or, you can rebuild ppl.nn without JIT, then test again:
cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF
from ppl.nn.
--quick-select
Using --quick-select can fix the problem. But the profiling performance is poor. Running mobilenet v2 with ppl.nn using default kernel costs 0.78ms, while running with tensorrt costs 0.39ms.
Can I use kernel tuning on mobilenet v2?
from ppl.nn.
If you consider about the perfermance, you can rebuild ppl.nn without JIT. I think this way can also fix the problem you met.
cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF
from ppl.nn.
If you consider about the perfermance, you can rebuild ppl.nn without JIT. I think this way can also fix the problem you met.
cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF
Thanks. The performance of ppl without JIT is great.
Do ppl only support fp16 on 2080Ti? I tried to set the kernel type as fp32 but failed.
./build-no-jit/tools/pplnn --use-cuda --onnx-model=resnet18.onnx --kernel-type=float32
[ERROR][2022-12-07 14:42:06.426][sequential_scheduler.cc:129] exec kernel[/maxpool/MaxPool] of type[:MaxPool:11] failed: unsupported
[ERROR][2022-12-07 14:42:06.426][runtime_impl.cc:337] Run() failed: unsupported
[ERROR][2022-12-07 14:42:06.426][pplnn.cc:1315] Run() failed: unsupported
from ppl.nn.
We only support fp16/int8 kernels for conv right now. So, the runtime crashed before conv.
from ppl.nn.
We only support fp16/int8 kernels for conv right now. So, the runtime crashed before conv.
OK. Thanks for your reply.
from ppl.nn.
Do you mind sent your broken model to [email protected]
from ppl.nn.
Related Issues (20)
- Onnx run error HOT 2
- 请问支持int8在高通芯片上cDSP进行推理吗?
- Slice op question HOT 1
- linux compile error protobuf static assertion failed HOT 3
- malloc_consolidate(): invalid chunk size HOT 2
- pplnn save-input 得到的NDARRAY的 shape不正确 HOT 1
- 如何使用cmake的将ppl.nn和依赖ppl.nn的代码一同编译? HOT 3
- Segmentation fault at ppl::nn::x86::X86Kernel::DumpOutputTensors HOT 5
- 获取模型推理结果(GetOutputs)耗时长 HOT 2
- Install Error HOT 1
- The compilation passed, but an error was reported in test phase HOT 2
- Floating point exception (core dumped) ? HOT 4
- 使用x86 engine运行resnet50 fp16 onnx模型 core dump
- (Ask) why InferInheritedType handle int8 to fp16 out? HOT 3
- Got wrong output shape when run a Gemm op(transB=0) use cuda HOT 4
- Crash with ONNX Split operator
- 关于全局engine,其他线程引用导致的性能下降问题 HOT 4
- 推理误差排查
- 多模型pipeline的示例
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from ppl.nn.