transformers
fbb2054e - Offloaded hybrid cache for Llama4 (#37401)

Commit
248 days ago
Offloaded hybrid cache for Llama4 (#37401) * first try (maybe race condition) * Update cache_utils.py * cannot avoid the race condition -> use 2 layers * Update cache_utils.py * Update cache_utils.py
Author
Parents
Loading