commit 67b2fbad227eba67a4f01376117e0cf1acdfc30a
parent ddb4bfba50cb563324e687403164ecc718ea1211
Author: serge-sans-paille <sguelton@mozilla.com>
Date: Tue, 16 Dec 2025 19:00:17 +0000
Bug 2006241 - Fix llama implicit includes r=cpeterson,padenot
Backported from https://github.com/ggml-org/llama.cpp/commit/2aa45ef9e31069ea0a7d0fef7ce858facdf25218
Differential Revision: https://phabricator.services.mozilla.com/D276618
Diffstat:
3 files changed, 16 insertions(+), 0 deletions(-)
diff --git a/third_party/llama.cpp/missing-include.patch b/third_party/llama.cpp/missing-include.patch
@@ -0,0 +1,13 @@
+diff --git a/src/llama-hparams.cpp b/src/llama-hparams.cpp
+index 7a06368dcda68..180d0c81e62cc 100644
+--- a/src/llama-hparams.cpp
++++ b/src/llama-hparams.cpp
+@@ -2,6 +2,8 @@
+
+ #include "ggml.h"
+
++#include <algorithm>
++
+ void llama_hparams::set_swa_pattern(uint32_t n_pattern, bool dense_first) {
+ if (dense_first) {
+ for (uint32_t il = 0; il < n_layer; ++il) {
diff --git a/third_party/llama.cpp/moz.yaml b/third_party/llama.cpp/moz.yaml
@@ -84,3 +84,4 @@ vendoring:
- load-from-buffer-or-fd.patch # add an API to load a model from a buffer or a fd
- static-ctors.patch # remove some static ctors
- fix-RLIMIT_MEMLOCK-solaris.patch # fix the build on solaris
+ - missing-include.patch # backported from https://github.com/ggml-org/llama.cpp/commit/2aa45ef9e31069ea0a7d0fef7ce858facdf25218
diff --git a/third_party/llama.cpp/src/llama-hparams.cpp b/third_party/llama.cpp/src/llama-hparams.cpp
@@ -2,6 +2,8 @@
#include "ggml.h"
+#include <algorithm>
+
void llama_hparams::set_swa_pattern(uint32_t n_pattern, bool dense_first) {
if (dense_first) {
for (uint32_t il = 0; il < n_layer; ++il) {