llama.cpp
77d5e9a7 - ggml: move fp16/bf16 conversion optimizations to CPU backend + export conversion APIs (#13107)

Commit
134 days ago
ggml: move fp16/bf16 conversion optimizations to CPU backend + export conversion APIs (#13107) * ggml: dynamic x86_64 feature detection for FP32 <-> FP16/BF16 conversion * move fp converter to ggml-cpu * Switch ggml_compute_forward_get_rows_f16/bf16 to new ggml_cpu_fp16/bf16_to_fp32
Author
Parents
Loading