llama.cpp
f125b8dc - llama : add PLM GGUF Conversion & Inference Support (#12457)

Commit
261 days ago
llama : add PLM GGUF Conversion & Inference Support (#12457) * add edgellm model arch[conversation feature doesn't work] * remove output.weight layer for edgellm arch * [Model] update the name of the model * update the name of model arch in convert gguf * [Model] Refarctor the model arch into llama-model * [Bug] Fix the bug in create attn kv * [Code] Fix editorconfig erros * [Code] Remove Trailing whitespace * [Code] Remove Trailing whitespace * [Code] Change the order of model arch in list * [Code] Fix flake8 Lint errors * Remove trailing white space * [Code] Remove call in model arch
Author
Parents
Loading