Skip to content

Pull requests: ggml-org/llama.cpp

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

mtmd: add mtmd_context_params::warmup option
#17652 opened Dec 1, 2025 by ngxson Loading…
ggml: added missing cast sections in memcpy
#17651 opened Dec 1, 2025 by GermanAizek Loading…
ggml-cpu: remove duplicate conditional check 'iid'
#17650 opened Dec 1, 2025 by GermanAizek Loading…
sgemm: reuse loaded vector in AVX dot product calculation ggml changes relating to the ggml tensor library for machine learning
#17648 opened Dec 1, 2025 by GermanAizek Loading…
vec: optimize AVX2/FMA sum-of-squares with loop unrolling and FMA ggml changes relating to the ggml tensor library for machine learning
#17642 opened Dec 1, 2025 by GermanAizek Loading…
ggml-quants: use _mm256_testz_si256 for mask checks in AVX2 ggml changes relating to the ggml tensor library for machine learning
#17641 opened Dec 1, 2025 by GermanAizek Loading…
ggml-alloc: optimize free block shifting with memmove ggml changes relating to the ggml tensor library for machine learning
#17640 opened Dec 1, 2025 by GermanAizek Loading…
ggml-cuda: reorder only relevant nodes ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs
#17639 opened Dec 1, 2025 by am17an Loading…
vulkan: Replace deprecated VK_EXT_validation_features ggml changes relating to the ggml tensor library for machine learning Vulkan Issues specific to the Vulkan backend
#17637 opened Dec 1, 2025 by rillomas Loading…
common: improve verbosity level definitions
#17630 opened Nov 30, 2025 by ngxson Loading…
llama-router, the C++ "llama-swap" for llama.cpp examples need feedback Testing and feedback with results are needed testing Everything test related
#17629 opened Nov 30, 2025 by ServeurpersoCom Draft
vulkan: set all memory allocations to high priority ggml changes relating to the ggml tensor library for machine learning Vulkan Issues specific to the Vulkan backend
#17624 opened Nov 30, 2025 by jeffbolznv Draft
vulkan: Reduce temporary memory usage for TOP_K ggml changes relating to the ggml tensor library for machine learning Vulkan Issues specific to the Vulkan backend
#17623 opened Nov 30, 2025 by jeffbolznv Loading…
ggml : remove redundant n_copies check when setting input/output ggml changes relating to the ggml tensor library for machine learning
#17612 opened Nov 30, 2025 by danbev Loading…
common : add minimalist multi-thread progress bar
#17602 opened Nov 29, 2025 by angt Loading…
Feature/kimi linear support ggml changes relating to the ggml tensor library for machine learning model Model specific Nvidia GPU Issues specific to Nvidia GPUs python python script changes
#17592 opened Nov 29, 2025 by cacaview Loading…
Override SSM_A op for Qwen3 Next to reduce splits model Model specific
#17587 opened Nov 29, 2025 by pwilkin Loading…
Improve Qwen3-Next Speed model Model specific
#17585 opened Nov 29, 2025 by lovedheart Draft
Add support for CUMSUM and TRI for CUDA. ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs testing Everything test related
#17584 opened Nov 28, 2025 by pwilkin Loading…
Add safetensors support examples
#17580 opened Nov 28, 2025 by ericcurtin Loading…
ProTip! Mix and match filters to narrow down what you’re looking for.