llama.cpp
llama: use host memory if device reports 0 memory
#18587
Open

llama: use host memory if device reports 0 memory #18587

taronaeo wants to merge 4 commits into ggml-org:master from taronaeo:fix/ggml-backend-memory
taronaeo
taronaeo ggml: allow free = 0 and total = 0 to use host memory info
c53940ae
taronaeo taronaeo requested a review from lhez lhez 4 days ago
taronaeo taronaeo requested a review from max-krasnyansky max-krasnyansky 4 days ago
taronaeo taronaeo requested a review from ggerganov ggerganov 4 days ago
github-actions github-actions added ggml
github-actions github-actions added OpenCL
taronaeo taronaeo changed the title ggml: allow free = 0 and total = 0 to use host memory info ggml-backend: allow free = 0 and total = 0 to use host memory info 4 days ago
lhez
ggerganov
ggerganov commented on 2026-01-06
taronaeo ggml-backend: move memory fallback logic to llama_get_device_memory_data
84f09272
ggerganov
ggerganov commented on 2026-01-06
taronaeo llama: fix memory logic for devices with 0 free and total
00f8eb49
taronaeo ggml: revert new host memory api
05e8f735
taronaeo taronaeo requested a review from CISC CISC 1 day ago
CISC
CISC approved these changes on 2026-01-07
taronaeo taronaeo changed the title ggml-backend: allow free = 0 and total = 0 to use host memory info llama: use host memory if device reports 0 memory 8 hours ago
taronaeo

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone