llama.cpp
vulkan: Fix ErrorOutOfHostMemory on Intel GPU when loading large models with --no-mmap
#20059
Merged

vulkan: Fix ErrorOutOfHostMemory on Intel GPU when loading large models with --no-mmap #20059

rillomas
rillomas Changed to reuse command buffers to fix crashing on Intel GPU
4b52568b
github-actions github-actions added Vulkan
github-actions github-actions added ggml
rillomas Removed unused parameter
d1dd8147
HumerousGorgon
rillomas
rillomas Fixed compile error and minor mistake
668d245e
rillomas Fix logging
29a1a01a
rillomas rillomas marked this pull request as ready for review 20 days ago
rillomas rillomas requested a review from 0cc4m 0cc4m 20 days ago
0cc4m
0cc4m commented on 2026-03-06
danielmayost
rillomas rillomas marked this pull request as draft 15 days ago
rillomas Changing to use usage flag per command buffer
e1f8ce0c
rillomas
rillomas fixed style
19d54833
rillomas added buffer reset
ffed7e53
rillomas Removed cmd_buffer_idx for reuse consistency
a0fecda9
rillomas Merge remote-tracking branch 'origin/master' into fix-async-tensor-crash
498ff284
rillomas rillomas marked this pull request as ready for review 14 days ago
rillomas Fixed style
d3fab849
0cc4m
0cc4m approved these changes on 2026-03-12
0cc4m 0cc4m merged 5866e3bb into master 12 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone