llama.cpp
Make loading weights 10-100x faster
#613
Merged

Make loading weights 10-100x faster #613

jart merged 9 commits into ggml-org:master from jart:loader
jart
slaren Add mmap support for model files
2a6cef62
slaren Fix ggml_init_params in quantize
a1e0f17a
slaren Make mmap_file static
4ae12d08
slaren Unmap the file in llama_free
4daaa5e7
slaren Always initialize mm_addr and mm_length in llama_model
812cfa19
slaren Initial windows support (untested)
80c2178d
jart jart added performance
jart jart added breaking change
luminalle
jart jart force pushed to b806987b 2 years ago
jart
jart jart force pushed from b806987b 2 years ago
jart
FNsi
bakkot
bakkot commented on 2023-03-30
ggerganov
jart Make loading weights 10-100x faster
75d1e551
jart jart force pushed to 75d1e551 2 years ago
jart
jart Ensure --mlock works properly with mmap() support
a45e843e
mqy
mqy commented on 2023-03-30
sw
sw commented on 2023-03-30
jart jart force pushed 2 years ago
jart
x02Sylvie
jart
x02Sylvie
jart
rabidcopy
jart jart force pushed to adaba698 2 years ago
jart
ggerganov
ggerganov approved these changes on 2023-03-30
CoderRC
CoderRC
jart Introduce GGML migration tool for new file format
516474b4
jart jart force pushed from adaba698 to 516474b4 2 years ago
jart
jart jart merged ee0c40dd into master 2 years ago
gaceladri
gaceladri
rabidcopy
gaceladri
asklar

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone