--- /srv/rebuilderd/tmp/rebuilderdvN7gMY/inputs/llama.cpp-tools-extra_8941+dfsg-1_arm64.deb +++ /srv/rebuilderd/tmp/rebuilderdvN7gMY/out/llama.cpp-tools-extra_8941+dfsg-1_arm64.deb ├── file list │ @@ -1,3 +1,3 @@ │ -rw-r--r-- 0 0 0 4 2026-04-26 21:45:18.000000 debian-binary │ -rw-r--r-- 0 0 0 1876 2026-04-26 21:45:18.000000 control.tar.xz │ --rw-r--r-- 0 0 0 339612 2026-04-26 21:45:18.000000 data.tar.xz │ +-rw-r--r-- 0 0 0 339684 2026-04-26 21:45:18.000000 data.tar.xz ├── control.tar.xz │ ├── control.tar │ │ ├── ./md5sums │ │ │ ├── ./md5sums │ │ │ │┄ Files differ ├── data.tar.xz │ ├── data.tar │ │ ├── file list │ │ │ @@ -26,15 +26,15 @@ │ │ │ -rw-r--r-- 0 root (0) root (0) 3442 2026-04-26 21:45:18.000000 ./usr/share/doc/llama.cpp-tools-extra/changelog.Debian.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 13361 2026-04-26 21:45:18.000000 ./usr/share/doc/llama.cpp-tools-extra/copyright │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-26 21:45:18.000000 ./usr/share/lintian/ │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-26 21:45:18.000000 ./usr/share/lintian/overrides/ │ │ │ -rw-r--r-- 0 root (0) root (0) 180 2026-04-26 21:45:18.000000 ./usr/share/lintian/overrides/llama.cpp-tools-extra │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-26 21:45:18.000000 ./usr/share/man/ │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-26 21:45:18.000000 ./usr/share/man/man1/ │ │ │ --rw-r--r-- 0 root (0) root (0) 6393 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-batched-bench.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 6398 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-batched-bench.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 514 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-gguf-split.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 6585 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-imatrix.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 7142 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-mtmd-cli.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 6546 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-perplexity.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 6553 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-perplexity.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 6130 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-results.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 728 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-tokenize.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 6248 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-tts.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 6249 2026-04-26 21:45:18.000000 ./usr/share/man/man1/llama-tts.1.gz │ │ ├── ./usr/share/man/man1/llama-batched-bench.1.gz │ │ │ ├── llama-batched-bench.1 │ │ │ │ @@ -1,13 +1,12 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-BATCHED-BENCH "1" "April 2026" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-batched-bench \- llama-batched-bench │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-license\fR show source code license and dependencies │ │ │ │ \fB\-cl\fR, \fB\-\-cache\-list\fR show list of models in cache │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ @@ -412,16 +411,19 @@ │ │ │ │ \fB\-\-dry\-base\fR N set DRY sampling base value (default: 1.75) │ │ │ │ \fB\-\-dry\-allowed\-length\fR N set allowed length for DRY sampling (default: 2) │ │ │ │ \fB\-\-dry\-penalty\-last\-n\fR N set DRY penalty for the last n tokens (default: \fB\-1\fR, 0 = disable, \fB\-1\fR = │ │ │ │ .IP │ │ │ │ context size) │ │ │ │ .PP │ │ │ │ \fB\-\-dry\-sequence\-breaker\fR STRING add sequence breaker for DRY sampling, clearing out default breakers │ │ │ │ -.TP │ │ │ │ -('\en', ':', '"', '*') in the process; use "none" to not use any │ │ │ │ +.IP │ │ │ │ +('\en', ':', '"', '*') in the process; uload_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ +.PP │ │ │ │ +se "none" to not use any │ │ │ │ +.IP │ │ │ │ sequence breakers │ │ │ │ .PP │ │ │ │ \fB\-\-adaptive\-target\fR N adaptive\-p: select tokens near this probability (valid range 0.0 to │ │ │ │ .TP │ │ │ │ 1.0; negative = disabled) (default: \fB\-1\fR.00) │ │ │ │ [(more info)](https://github.com/ggml\-org/llama.cpp/pull/17927) │ │ │ │ .PP │ │ ├── ./usr/share/man/man1/llama-perplexity.1.gz │ │ │ ├── llama-perplexity.1 │ │ │ │ @@ -1,13 +1,12 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-PERPLEXITY "1" "April 2026" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-perplexity \- llama-perplexity │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-license\fR show source code license and dependencies │ │ │ │ \fB\-cl\fR, \fB\-\-cache\-list\fR show list of models in cache │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ @@ -489,15 +488,16 @@ │ │ │ │ \fB\-\-winogrande\fR compute Winogrande score over random tasks from datafile supplied with │ │ │ │ .IP │ │ │ │ \fB\-f\fR │ │ │ │ .SS "--winogrande-tasks N number of tasks to use when computing the Winogrande score (default:" │ │ │ │ .IP │ │ │ │ 0) │ │ │ │ .PP │ │ │ │ -\fB\-\-multiple\-choice\fR compute multiple choice score over random tasks from datafile supplied │ │ │ │ +\fB\-\-mload_backend\fR: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ +ultiple\-choice compute multiple choice score over random tasks from datafile supplied │ │ │ │ .IP │ │ │ │ with \fB\-f\fR │ │ │ │ .PP │ │ │ │ \fB\-\-multiple\-choice\-tasks\fR N number of tasks to use when computing the multiple choice score │ │ │ │ .IP │ │ │ │ (default: 0) │ │ │ │ .PP │ │ ├── ./usr/share/man/man1/llama-tts.1.gz │ │ │ ├── llama-tts.1 │ │ │ │ @@ -1,13 +1,12 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-TTS "1" "April 2026" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-tts \- llama-tts │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-license\fR show source code license and dependencies │ │ │ │ \fB\-cl\fR, \fB\-\-cache\-list\fR show list of models in cache │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ @@ -414,15 +413,16 @@ │ │ │ │ context size) │ │ │ │ .PP │ │ │ │ \fB\-\-dry\-sequence\-breaker\fR STRING add sequence breaker for DRY sampling, clearing out default breakers │ │ │ │ .TP │ │ │ │ ('\en', ':', '"', '*') in the process; use "none" to not use any │ │ │ │ sequence breakers │ │ │ │ .PP │ │ │ │ -\fB\-\-adaptive\-target\fR N adaptive\-p: select tokens near this probability (valid range 0.0 to │ │ │ │ +\fB\-\-adaptive\-target\fR N adaptive\-p: select toload_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ +kens near this probability (valid range 0.0 to │ │ │ │ .TP │ │ │ │ 1.0; negative = disabled) (default: \fB\-1\fR.00) │ │ │ │ [(more info)](https://github.com/ggml\-org/llama.cpp/pull/17927) │ │ │ │ .PP │ │ │ │ \fB\-\-adaptive\-decay\fR N adaptive\-p: decay rate for target adaptation over time. lower values │ │ │ │ .TP │ │ │ │ are more reactive, higher values are more stable.