Unknown argument: --origin-rpc-server-main-gpu [2025-07-18 09:19:49.589966] I [2025-07-18 09:19:49.589966] I arguments : /usr/local/lib/python3.10/site-packages/gpustack/third_party/bin/llama-box/llama-box-rpc-server --rpc-server-host 0.0.0.0 --rpc-server-port 40093 --rpc-server-main-gpu 0 --origin-rpc-server-main-gpu 7 [2025-07-18 09:19:49.589966] I version : v0.0.154 (53fe21f) [2025-07-18 09:19:49.589966] I compiler : cc (Ubuntu 11.4.0-2ubuntu1~18.04) 11.4.0 [2025-07-18 09:19:49.589966] I target : aarch64-linux-gnu [2025-07-18 09:19:49.589966] I vendor : llama.cpp 3ac67535 (5586), stable-diffusion.cpp 3eb18db (204), concurrentqueue 2f09da7 (295), readerwriterqueue 16b48ae (166) [2025-07-18 09:19:49.589968] I system_info: n_threads = 128 (n_threads_batch = 128) / 128 | CPU : NEON = 1 | ARM_FMA = 1 | LLAMAFILE = 1 | AARCH64_REPACK = 1 | [2025-07-18 09:19:49.589968] I [2025-07-18 09:19:49.589970] I srv load: using GPU backend [2025-07-18 09:19:49.589970] I srv rpcserver_create_backend: fallback, using CPU backend [2025-07-18 09:19:49.589970] I srv start: starting [2025-07-18 09:19:49.589973] I srv start: proto v2.0.0, listening host = 0.0.0.0, port = 40093, capacity_mib = 522228