We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
# ./llama-cli --version version: 5218 (00e3e5a1) built with cc (Debian 12.2.0-14) 12.2.0 for x86_64-linux-gnu
Linux
No response
If rpc-server is started as a background process and no cache is specified, it fails.
rpc-server
root@debian12-dev:~# systemctl status lxllamad.service ● lxllamad.service - Lxllamad service for distributed inference Loaded: loaded (/lib/systemd/system/lxllamad.service; enabled; preset: enabled) Active: active (running) since Tue 2025-04-29 10:00:22 EDT; 2s ago Main PID: 54469 (lxllamad) Tasks: 4 (limit: 9442) Memory: 3.0M CPU: 14ms CGroup: /system.slice/lxllamad.service └─54469 /opt/lxllama/lxllamad Apr 29 10:00:22 debian12-dev systemd[1]: Started lxllamad.service - Lxllamad service for distributed inference. Apr 29 10:00:22 debian12-dev lxllamad[54491]: !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! Apr 29 10:00:22 debian12-dev lxllamad[54491]: WARNING: Host ('0.0.0.0') is != '127.0.0.1' Apr 29 10:00:22 debian12-dev lxllamad[54491]: Never expose the RPC server to an open network! Apr 29 10:00:22 debian12-dev lxllamad[54491]: This is an experimental feature and is not secure! Apr 29 10:00:22 debian12-dev lxllamad[54491]: !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! Apr 29 10:00:22 debian12-dev lxllamad[54491]: create_backend: using CPU backend root@debian12-dev:~# ps -efl | grep rpc-server 0 Z root 54491 54469 0 80 0 - 0 - 10:00 ? 00:00:00 [rpc-server] <defunct> 0 S root 54495 1036 0 80 0 - 1584 pipe_r 10:00 pts/1 00:00:00 grep --color=auto rpc-server
rpc-server is started like this: /opt/llama/bin/rpc-server -t 8
The text was updated successfully, but these errors were encountered:
Was this fixed in #13188?
Sorry, something went wrong.
No branches or pull requests
Name and Version
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
No response
Command line
Problem description & steps to reproduce
If
rpc-server
is started as a background process and no cache is specified, it fails.rpc-server
is started like this: /opt/llama/bin/rpc-server -t 8First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: