llama.cpp
CUDA: fix MMQ writeback for int8 tensor cores
#8100
Merged

CUDA: fix MMQ writeback for int8 tensor cores #8100

JohannesGaessler
JohannesGaessler CUDA: fix MMQ writeback for int8 tensor cores
0402d4f8
github-actions github-actions added Nvidia GPU
github-actions github-actions added ggml
slaren
slaren approved these changes on 2024-06-24
JohannesGaessler JohannesGaessler merged 3b099bcd into master 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone