llama.cpp
9a5c2a16 - cuda : switch to F16 scalars + tune warps for RTX 2060

Commit
1 year ago
cuda : switch to F16 scalars + tune warps for RTX 2060
Author
Committer
Parents
  • File
    ggml-cuda.cu
  • tests
    • File
      test-backend-ops.cpp