Git Product home page Git Product logo

Comments (7)

Si-XU avatar Si-XU commented on June 18, 2024

Cause we donot have 2080Ti right now, we cannot reproduce your error locally. There are two ways to skip such error I used.

You can firstly test if it still get error by using default kernel.

./pplnn-build/tools/pplnn --use-cuda --onnx-model=mobilenet_v2.onnx --kernel-type=float16 --quick-select

Or, you can rebuild ppl.nn without JIT, then test again:

cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF

from ppl.nn.

shiwenloong avatar shiwenloong commented on June 18, 2024

--quick-select

Using --quick-select can fix the problem. But the profiling performance is poor. Running mobilenet v2 with ppl.nn using default kernel costs 0.78ms, while running with tensorrt costs 0.39ms.

Can I use kernel tuning on mobilenet v2?

from ppl.nn.

Si-XU avatar Si-XU commented on June 18, 2024

If you consider about the perfermance, you can rebuild ppl.nn without JIT. I think this way can also fix the problem you met.

cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF

from ppl.nn.

shiwenloong avatar shiwenloong commented on June 18, 2024

If you consider about the perfermance, you can rebuild ppl.nn without JIT. I think this way can also fix the problem you met.

cmake .. -DCMAKE_BUILD_TYPE=Release -DPPLNN_USE_CUDA=ON -DCMAKE_INSTALL_PREFIX=install -DPPLNN_ENABLE_CUDA_JIT=OFF

Thanks. The performance of ppl without JIT is great.
Do ppl only support fp16 on 2080Ti? I tried to set the kernel type as fp32 but failed.

./build-no-jit/tools/pplnn  --use-cuda --onnx-model=resnet18.onnx --kernel-type=float32
[ERROR][2022-12-07 14:42:06.426][sequential_scheduler.cc:129] exec kernel[/maxpool/MaxPool] of type[:MaxPool:11] failed: unsupported
[ERROR][2022-12-07 14:42:06.426][runtime_impl.cc:337] Run() failed: unsupported
[ERROR][2022-12-07 14:42:06.426][pplnn.cc:1315] Run() failed: unsupported

from ppl.nn.

Si-XU avatar Si-XU commented on June 18, 2024

We only support fp16/int8 kernels for conv right now. So, the runtime crashed before conv.

from ppl.nn.

shiwenloong avatar shiwenloong commented on June 18, 2024

We only support fp16/int8 kernels for conv right now. So, the runtime crashed before conv.

OK. Thanks for your reply.

from ppl.nn.

Si-XU avatar Si-XU commented on June 18, 2024

Do you mind sent your broken model to [email protected]

from ppl.nn.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.