llama.cpp
Fix conversion of unnormalized BF16->BF16 weights
#7843
Merged

Fix conversion of unnormalized BF16->BF16 weights #7843

compilade merged 10 commits into ggml-org:master from CISC:convert-bf16-fix
CISC
CISC add truncate_bf16
6a52bfe3
CISC truncate intermediate fp32 if converting bf16 to bf16
46054d1a
github-actions github-actions added python
mofosyne mofosyne added Review Complexity : Low
compilade
CISC
CISC
compilade
CISC
compilade
CISC
CISC
CISC fix masking in __compute_fp32_to_bf16
069369f3
CISC np.int16 no longer used
225ec48f
compilade
CISC
compilade
CISC
CISC
jart
jart
CISC missing cast and additional numpy 2.x fix
e8e2b7e0
CISC
CISC
jart
compilade ggml-impl : do not flush bf16 subnormals to zero
5b67a6cf
compilade
compilade Merge branch 'master' into convert-bf16-fix
675a7410
github-actions github-actions added ggml
compilade
jart
compilade
CISC Merge branch 'master' of github.com:ggerganov/llama.cpp into convert-…
2b746488
CISC missed prototype update in merge
dc051541
CISC merge cleanup
3a3a7528
CISC
compilade
compilade approved these changes on 2024-08-01
mofosyne
mofosyne commented on 2024-08-01
mofosyne mofosyne added merge ready
compilade compilade merged b72c20b8 into master 1 year ago
CISC CISC deleted the convert-bf16-fix branch 1 year ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone