--- /srv/rebuilderd/tmp/rebuilderdyqnARE/inputs/llama.cpp-tools_8681+dfsg-1_arm64.deb +++ /srv/rebuilderd/tmp/rebuilderdyqnARE/out/llama.cpp-tools_8681+dfsg-1_arm64.deb ├── file list │ @@ -1,3 +1,3 @@ │ -rw-r--r-- 0 0 0 4 2026-04-07 16:34:28.000000 debian-binary │ -rw-r--r-- 0 0 0 3440 2026-04-07 16:34:28.000000 control.tar.xz │ --rw-r--r-- 0 0 0 3338776 2026-04-07 16:34:28.000000 data.tar.xz │ +-rw-r--r-- 0 0 0 3338752 2026-04-07 16:34:28.000000 data.tar.xz ├── control.tar.xz │ ├── control.tar │ │ ├── ./md5sums │ │ │ ├── ./md5sums │ │ │ │┄ Files differ ├── data.tar.xz │ ├── data.tar │ │ ├── file list │ │ │ @@ -49,8 +49,8 @@ │ │ │ -rw-r--r-- 0 root (0) root (0) 259586 2026-04-06 18:54:06.000000 ./usr/share/llama.cpp-tools/llama-server/themes/wild/llamapattern.png │ │ │ -rw-r--r-- 0 root (0) root (0) 496463 2026-04-06 18:54:06.000000 ./usr/share/llama.cpp-tools/llama-server/themes/wild/wild.png │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-07 16:34:28.000000 ./usr/share/man/ │ │ │ drwxr-xr-x 0 root (0) root (0) 0 2026-04-07 16:34:28.000000 ./usr/share/man/man1/ │ │ │ -rw-r--r-- 0 root (0) root (0) 1515 2026-04-07 16:34:28.000000 ./usr/share/man/man1/llama-bench.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 8641 2026-04-07 16:34:28.000000 ./usr/share/man/man1/llama-cli.1.gz │ │ │ -rw-r--r-- 0 root (0) root (0) 1644 2026-04-07 16:34:28.000000 ./usr/share/man/man1/llama-quantize.1.gz │ │ │ --rw-r--r-- 0 root (0) root (0) 10484 2026-04-07 16:34:28.000000 ./usr/share/man/man1/llama-server.1.gz │ │ │ +-rw-r--r-- 0 root (0) root (0) 10480 2026-04-07 16:34:28.000000 ./usr/share/man/man1/llama-server.1.gz │ │ ├── ./usr/share/man/man1/llama-server.1.gz │ │ │ ├── llama-server.1 │ │ │ │ @@ -1,13 +1,12 @@ │ │ │ │ .\" DO NOT MODIFY THIS FILE! It was generated by help2man 1.49.3. │ │ │ │ .TH LLAMA-SERVER "1" "April 2026" "debian" "User Commands" │ │ │ │ .SH NAME │ │ │ │ llama-server \- llama-server │ │ │ │ .SH DESCRIPTION │ │ │ │ -load_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ \fB\-\-\-\-\-\fR common params \fB\-\-\-\-\-\fR │ │ │ │ .PP │ │ │ │ \fB\-h\fR, \fB\-\-help\fR, \fB\-\-usage\fR print usage and exit │ │ │ │ \fB\-\-version\fR show version and build info │ │ │ │ \fB\-\-license\fR show source code license and dependencies │ │ │ │ \fB\-cl\fR, \fB\-\-cache\-list\fR show list of models in cache │ │ │ │ \fB\-\-completion\-bash\fR print source\-able bash completion script for llama.cpp │ │ │ │ @@ -83,15 +82,16 @@ │ │ │ │ the model │ │ │ │ (env: LLAMA_ARG_ROPE_SCALING_TYPE) │ │ │ │ .PP │ │ │ │ \fB\-\-rope\-scale\fR N RoPE context scaling factor, expands context by a factor of N │ │ │ │ .IP │ │ │ │ (env: LLAMA_ARG_ROPE_SCALE) │ │ │ │ .PP │ │ │ │ -\fB\-\-rope\-freq\-base\fR N RoPE base frequency, used by NTK\-aware scaling (default: loaded from │ │ │ │ +\fB\-\-rope\-freq\-base\fR N RoPE base frequency, used by load_backend: loaded CPU backend from \fI\,/usr/lib/aarch64\-linux\-gnu/ggml/backends0/libggml\-cpu\-armv8.2_2.so\/\fP │ │ │ │ +NTK\-aware scaling (default: loaded from │ │ │ │ .TP │ │ │ │ model) │ │ │ │ (env: LLAMA_ARG_ROPE_FREQ_BASE) │ │ │ │ .PP │ │ │ │ \fB\-\-rope\-freq\-scale\fR N RoPE frequency scaling factor, expands context by a factor of 1/N │ │ │ │ .IP │ │ │ │ (env: LLAMA_ARG_ROPE_FREQ_SCALE)