--- /srv/rebuilderd/tmp/rebuilderd8dHauj/inputs/llama.cpp-examples_6641+dfsg-1_ppc64el.deb +++ /srv/rebuilderd/tmp/rebuilderd8dHauj/out/llama.cpp-examples_6641+dfsg-1_ppc64el.deb ├── file list │ @@ -1,3 +1,3 @@ │ -rw-r--r-- 0 0 0 4 2025-10-05 20:09:48.000000 debian-binary │ --rw-r--r-- 0 0 0 2320 2025-10-05 20:09:48.000000 control.tar.xz │ --rw-r--r-- 0 0 0 2818588 2025-10-05 20:09:48.000000 data.tar.xz │ +-rw-r--r-- 0 0 0 2324 2025-10-05 20:09:48.000000 control.tar.xz │ +-rw-r--r-- 0 0 0 2820140 2025-10-05 20:09:48.000000 data.tar.xz ├── control.tar.xz │ ├── control.tar │ │ ├── ./md5sums │ │ │ ├── ./md5sums │ │ │ │┄ Files differ ├── data.tar.xz │ ├── data.tar │ │ ├── file list │ │ │ @@ -29,26 +29,26 @@ │ │ │ -rw-r--r-- 0 root (0) root (0) 2019 2025-10-05 20:09:48.000000 ./usr/share/doc/llama.cpp-examples/changelog.Debian.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 13060 2025-10-05 20:09:48.000000 ./usr/share/doc/llama.cpp-examples/copyright │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2025-10-05 20:09:48.000000 ./usr/share/lintian/ │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2025-10-05 20:09:48.000000 ./usr/share/lintian/overrides/ │ │ │ -rw-r--r-- 0 root (0) root (0) 270 2025-10-05 20:09:48.000000 ./usr/share/lintian/overrides/llama.cpp-examples │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2025-10-05 20:09:48.000000 ./usr/share/man/ │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2025-10-05 20:09:48.000000 ./usr/share/man/man1/ │ │ │ --rw-r--r-- 0 root (0) root (0) 5537 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-batched.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5942 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-embedding.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5489 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-eval-callback.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5538 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-batched.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5943 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-embedding.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5490 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-eval-callback.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5831 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-finetune.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 402 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-gguf-hash.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 256 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-gguf.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5485 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookahead.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5701 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup-create.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5702 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup-create.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 275 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup-merge.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5699 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup-stats.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5691 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5559 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-parallel.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5645 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-passkey.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5700 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup-stats.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5692 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-lookup.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5560 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-parallel.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5646 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-passkey.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 5782 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-retrieval.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 5493 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-save-load-state.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 5494 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-save-load-state.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 273 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-simple-chat.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 264 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-simple.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 6176 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-speculative-simple.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 6167 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-speculative.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 6177 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-speculative-simple.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 6166 2025-10-05 20:09:48.000000 ./usr/share/man/man1/llama-speculative.1.gz │ │ ├── ./usr/share/man/man1/llama-batched.1.gz │ │ │ ├── llama-batched.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-BATCHED "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-batched \- llama-batched │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-embedding.1.gz │ │ │ ├── llama-embedding.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-EMBEDDING "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-embedding \- llama-embedding │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-eval-callback.1.gz │ │ │ ├── llama-eval-callback.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-EVAL-CALLBACK "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-eval-callback \- llama-eval-callback │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-finetune.1.gz │ │ │ ├── llama-finetune.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-FINETUNE "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-finetune \- llama-finetune │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-lookahead.1.gz │ │ │ ├── llama-lookahead.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKAHEAD "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookahead \- llama-lookahead │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-lookup-create.1.gz │ │ │ ├── llama-lookup-create.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP-CREATE "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup-create \- llama-lookup-create │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-lookup-stats.1.gz │ │ │ ├── llama-lookup-stats.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP-STATS "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup-stats \- llama-lookup-stats │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-lookup.1.gz │ │ │ ├── llama-lookup.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-LOOKUP "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-lookup \- llama-lookup │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-parallel.1.gz │ │ │ ├── llama-parallel.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-PARALLEL "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-parallel \- llama-parallel │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-passkey.1.gz │ │ │ ├── llama-passkey.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-PASSKEY "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-passkey \- llama-passkey │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-retrieval.1.gz │ │ │ ├── llama-retrieval.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-RETRIEVAL "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-retrieval \- llama-retrieval │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-save-load-state.1.gz │ │ │ ├── llama-save-load-state.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SAVE-LOAD-STATE "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-save-load-state \- llama-save-load-state │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-\-\-\-\fR sampling params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-\-samplers\fR SAMPLERS samplers that will be used for generation in the order, separated by │ │ │ │ .TP │ │ │ │ \&';' │ │ ├── ./usr/share/man/man1/llama-speculative-simple.1.gz │ │ │ ├── llama-speculative-simple.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SPECULATIVE-SIMPLE "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-speculative-simple \- llama-speculative-simple │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR) │ │ ├── ./usr/share/man/man1/llama-speculative.1.gz │ │ │ ├── llama-speculative.1 │ │ │ │ @@ -1,13 +1,13 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SPECULATIVE "1" "October 2025" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-speculative \- llama-speculative │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power9.so\/\fP │ │ │ │ +load_backend: loaded CPU backend from \fI\,/usr/lib/powerpc64le\-linux\-gnu/ggml/backends0/libggml\-cpu\-power10.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ \fB\-\-verbose\-prompt\fR print a verbose prompt before generation (default: false) │ │ │ │ \fB\-t\fR, \fB\-\-threads\fR N number of threads to use during generation (default: \fB\-1\fR)