llama.cpp
f66f5829 - llama : refactor `src/llama.cpp` (#10902)

Comment changes are shownComment changes are hidden
Commit
189 days ago
llama : refactor `src/llama.cpp` (#10902) * llama : scatter llama.cpp into multiple modules (wip) * llama : control-vector -> adapter * llama : arch * llama : mmap ggml-ci * ci : remove BUILD_SHARED_LIBS=OFF ggml-ci * llama : arch (cont) ggml-ci * llama : chat ggml-ci * llama : model ggml-ci * llama : hparams ggml-ci * llama : adapter ggml-ci * examples : fix ggml-ci * rebase ggml-ci * minor * llama : kv cache ggml-ci * llama : impl ggml-ci * llama : batch ggml-ci * cont ggml-ci * llama : context ggml-ci * minor * llama : context (cont) ggml-ci * llama : model loader ggml-ci * common : update lora ggml-ci * llama : quant ggml-ci * llama : quant (cont) ggml-ci * minor [no ci]
Author
Parents
  • .github/workflows
    • File
      build.yml
  • common
    • File
      arg.cpp
    • File
      common.cpp
    • File
      common.h
  • examples
    • convert-llama2c-to-ggml
      • File
        convert-llama2c-to-ggml.cpp
    • cvector-generator
      • File
        cvector-generator.cpp
    • embedding
      • File
        embedding.cpp
    • eval-callback
      • File
        eval-callback.cpp
    • gguf-split
      • File
        gguf-split.cpp
    • imatrix
      • File
        imatrix.cpp
    • infill
      • File
        infill.cpp
    • lookahead
      • File
        lookahead.cpp
    • lookup
      • File
        lookup-create.cpp
      • File
        lookup-stats.cpp
      • File
        lookup.cpp
    • main
      • File
        main.cpp
    • parallel
      • File
        parallel.cpp
    • perplexity
      • File
        perplexity.cpp
    • quantize-stats
      • File
        quantize-stats.cpp
    • retrieval
      • File
        retrieval.cpp
    • save-load-state
      • File
        save-load-state.cpp
    • server
      • File
        server.cpp
      • File
        utils.hpp
    • speculative-simple
      • File
        speculative-simple.cpp
    • speculative
      • File
        speculative.cpp
    • tts
      • File
        tts.cpp
  • include
    • File
      llama-cpp.h
    • File
      llama.h
  • src
    • File
      CMakeLists.txt
    • File
      llama-adapter.cpp
    • File
      llama-adapter.h
    • File
      llama-arch.cpp
    • File
      llama-arch.h
    • File
      llama-batch.cpp
    • File
      llama-batch.h
    • File
      llama-chat.cpp
    • File
      llama-chat.h
    • File
      llama-context.cpp
    • File
      llama-context.h
    • File
      llama-cparams.cpp
    • File
      llama-cparams.h
    • File
      llama-grammar.cpp
    • File
      llama-grammar.h
    • File
      llama-hparams.cpp
    • File
      llama-hparams.h
    • File
      llama-impl.cpp
    • File
      llama-impl.h
    • File
      llama-kv-cache.cpp
    • File
      llama-kv-cache.h
    • File
      llama-mmap.cpp
    • File
      llama-mmap.h
    • File
      llama-model-loader.cpp
    • File
      llama-model-loader.h
    • File
      llama-model.cpp
    • File
      llama-model.h
    • File
      llama-quant.cpp
    • File
      llama-quant.h
    • File
      llama-sampling.cpp
    • File
      llama-vocab.cpp
    • File
      llama-vocab.h
    • File
      llama.cpp