llama.cpp
llama/ggml: multi-GPU pipeline parallelism (xdev host staging) + faster model loading
#19922
Closed
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
llama/ggml: multi-GPU pipeline parallelism (xdev host staging) + faster model loading
#19922
mxxm-t
wants to merge 1 commit into
ggml-org:master
from
mxxm-t:pipeline-parallelism
pipeline-parallelism: xdev host staging + load-time toggles
dee003ea
mxxm-t
requested a review
from
CISC
9 days ago
mxxm-t
requested a review
from
ggerganov
9 days ago
github-actions
added
Nvidia GPU
github-actions
added
ggml
mxxm-t
closed this
4 days ago
Login to write a write a comment.
Login via GitHub
Reviewers
CISC
ggerganov
Assignees
No one assigned
Labels
Nvidia GPU
ggml
Milestone
No milestone
Login to write a write a comment.
Login via GitHub