llama.cpp
ggml webgpu: fix workgroup dispatch limit for large batch sizes
#19965
Merged

ggml webgpu: fix workgroup dispatch limit for large batch sizes #19965

abhijitramesh
abhijitramesh ggml-webgpu: fix workgroup dispatch limit for large batch sizes
b0169df5
abhijitramesh ggml-webgpu: add bounds checking for over-dispatched workgroups
11253444
abhijitramesh abhijitramesh requested a review from reeselevine reeselevine 10 days ago
github-actions github-actions added ggml
reeselevine
reeselevine commented on 2026-02-28
abhijitramesh ggml-webgpu: add back TODO for spliting large sizes into batches
b92eee7d
reeselevine Optimize 2d workgroup provisioning
57e9ed26
reeselevine Set some parameters that increase speed
023b5658
reeselevine
reeselevine approved these changes on 2026-03-03
reeselevine reeselevine merged 49a7564a into master 7 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone