--- /srv/rebuilderd/tmp/rebuilderdPUdKHc/inputs/llama.cpp-examples_5882+dfsg-3_amd64.deb +++ /srv/rebuilderd/tmp/rebuilderdPUdKHc/out/llama.cpp-examples_5882+dfsg-3_amd64.deb ├── file list │ @@ -1,3 +1,3 @@ │ -rw-r--r-- 0 0 0 4 2025-08-27 05:01:15.000000 debian-binary │ -rw-r--r-- 0 0 0 2324 2025-08-27 05:01:15.000000 control.tar.xz │ --rw-r--r-- 0 0 0 2932972 2025-08-27 05:01:15.000000 data.tar.xz │ +-rw-r--r-- 0 0 0 2932836 2025-08-27 05:01:15.000000 data.tar.xz ├── control.tar.xz │ ├── control.tar │ │ ├── ./md5sums │ │ │ ├── ./md5sums │ │ │ │┄ Files differ ├── data.tar.xz │ ├── data.tar │ │ ├── file list │ │ │ @@ -39,15 +39,15 @@ │ │ │ -rw-r--r-- 0 root (0) root (0) 398 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-gguf-hash.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 252 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-gguf.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5152 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-gritlm.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5158 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookahead.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5381 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookup-create.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 276 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookup-merge.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5380 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookup-stats.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5371 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookup.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5372 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-lookup.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5238 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-parallel.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5323 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-passkey.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5459 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-retrieval.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5168 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-save-load-state.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 268 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-simple-chat.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 259 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-simple.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5710 2025-08-27 05:01:15.000000 ./usr/share/man/man1/llama-speculative-simple.1.gz │ │ ├── ./usr/share/man/man1/llama-batched.1.gz │ │ │ ├── llama-batched.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-BATCHED "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-batched \- llama-batched │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-embedding.1.gz │ │ │ ├── llama-embedding.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-EMBEDDING "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-embedding \- llama-embedding │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-eval-callback.1.gz │ │ │ ├── llama-eval-callback.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-EVAL-CALLBACK "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-eval-callback \- llama-eval-callback │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-finetune.1.gz │ │ │ ├── llama-finetune.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-FINETUNE "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-finetune \- llama-finetune │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-gritlm.1.gz │ │ │ ├── llama-gritlm.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-GRITLM "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-gritlm \- llama-gritlm │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-lookahead.1.gz │ │ │ ├── llama-lookahead.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKAHEAD "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookahead \- llama-lookahead │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-lookup-create.1.gz │ │ │ ├── llama-lookup-create.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP-CREATE "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup-create \- llama-lookup-create │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-lookup-stats.1.gz │ │ │ ├── llama-lookup-stats.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP-STATS "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup-stats \- llama-lookup-stats │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-lookup.1.gz │ │ │ ├── llama-lookup.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup \- llama-lookup │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-parallel.1.gz │ │ │ ├── llama-parallel.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-PARALLEL "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-parallel \- llama-parallel │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-passkey.1.gz │ │ │ ├── llama-passkey.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-PASSKEY "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-passkey \- llama-passkey │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-retrieval.1.gz │ │ │ ├── llama-retrieval.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-RETRIEVAL "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-retrieval \- llama-retrieval │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-save-load-state.1.gz │ │ │ ├── llama-save-load-state.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SAVE-LOAD-STATE "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-save-load-state \- llama-save-load-state │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-speculative-simple.1.gz │ │ │ ├── llama-speculative-simple.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SPECULATIVE-SIMPLE "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-speculative-simple \- llama-speculative-simple │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-speculative.1.gz │ │ │ ├── llama-speculative.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SPECULATIVE "1" "August 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-speculative \- llama-speculative │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-icelake.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/x86_64\-linux\-gnu/ggml/backends0/libggml\-cpu\-haswell.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR)