You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
llama.cpp : add pipeline parallelism support #6017. Good news: seems high priority and will probably be ready soon. If this and the CUDA memory release bug correction is ready please add a quick intermediate release integration to LLamaSharp. This is important.
Tracking issue for thread safety in llama.cpp. The global inference lock can be removed once this is resolved.
ggml-org/llama.cpp#3960
The text was updated successfully, but these errors were encountered: