-
Notifications
You must be signed in to change notification settings - Fork 13.9k
Pull requests: ggml-org/llama.cpp
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
ggml: use 'exists( const std::filesystem::path&, std::error_code&)' instead of 'exists( const std::filesystem::path&)' to enhance robustness
#17653
opened Dec 1, 2025 by
flyinskyin2013
Loading…
ggml-cpu: remove duplicate conditional check 'iid'
#17650
opened Dec 1, 2025 by
GermanAizek
Loading…
gguf: llama: use
= default for trivial constructors and destructors
#17649
opened Dec 1, 2025 by
GermanAizek
Loading…
sgemm: reuse loaded vector in AVX dot product calculation
ggml
changes relating to the ggml tensor library for machine learning
#17648
opened Dec 1, 2025 by
GermanAizek
Loading…
llama-vocab: replace postfix with prefix increment for iterators
#17646
opened Dec 1, 2025 by
GermanAizek
Loading…
vec: optimize AVX2/FMA sum-of-squares with loop unrolling and FMA
ggml
changes relating to the ggml tensor library for machine learning
#17642
opened Dec 1, 2025 by
GermanAizek
Loading…
ggml-quants: use _mm256_testz_si256 for mask checks in AVX2
ggml
changes relating to the ggml tensor library for machine learning
#17641
opened Dec 1, 2025 by
GermanAizek
Loading…
ggml-alloc: optimize free block shifting with changes relating to the ggml tensor library for machine learning
memmove
ggml
#17640
opened Dec 1, 2025 by
GermanAizek
Loading…
ggml-cuda: reorder only relevant nodes
ggml
changes relating to the ggml tensor library for machine learning
Nvidia GPU
Issues specific to Nvidia GPUs
#17639
opened Dec 1, 2025 by
am17an
Loading…
vulkan: Replace deprecated VK_EXT_validation_features
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#17637
opened Dec 1, 2025 by
rillomas
Loading…
common : compute average token length from vocabulary
#17632
opened Dec 1, 2025 by
yifant-code
•
Draft
llama-router, the C++ "llama-swap" for llama.cpp
examples
need feedback
Testing and feedback with results are needed
testing
Everything test related
#17629
opened Nov 30, 2025 by
ServeurpersoCom
•
Draft
vulkan: set all memory allocations to high priority
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#17624
opened Nov 30, 2025 by
jeffbolznv
•
Draft
vulkan: Reduce temporary memory usage for TOP_K
ggml
changes relating to the ggml tensor library for machine learning
Vulkan
Issues specific to the Vulkan backend
#17623
opened Nov 30, 2025 by
jeffbolznv
Loading…
model : Fix marker placement for LFM2-VL in single turn llama-mtmd-cli
examples
#17616
opened Nov 30, 2025 by
tdakhran
Loading…
ggml : remove redundant n_copies check when setting input/output
ggml
changes relating to the ggml tensor library for machine learning
#17612
opened Nov 30, 2025 by
danbev
Loading…
Feature/kimi linear support
ggml
changes relating to the ggml tensor library for machine learning
model
Model specific
Nvidia GPU
Issues specific to Nvidia GPUs
python
python script changes
#17592
opened Nov 29, 2025 by
cacaview
Loading…
Override SSM_A op for Qwen3 Next to reduce splits
model
Model specific
#17587
opened Nov 29, 2025 by
pwilkin
Loading…
Add support for CUMSUM and TRI for CUDA.
ggml
changes relating to the ggml tensor library for machine learning
Nvidia GPU
Issues specific to Nvidia GPUs
testing
Everything test related
#17584
opened Nov 28, 2025 by
pwilkin
Loading…
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.