Description
Hey,
Help me out a bit here.. i'm i an 100% idiot or what is going on ?
--
LocalAI version:
LocalAI, Native, no docker
Environment, CPU architecture, OS, and Version:
VM on proxmox with 1070 TI passtrough and Mac M4, latest
Describe the bug
Clean install, load (any) model and get this:
"localai" error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:43369: connect: connection refused""
No matter how i set it up (havent tested docker yet) - i always come back to the same error.
I load it up, i choose a model, i goto chat and type something.. and get this:
May 08 12:19:18 localai local-ai[1333]: 12:19PM INF Trying to load the model 'qwen3-32b' with the backend '[llama-cpp llama-cpp-fallback bark-cpp piper silero-vad stablediffusion-ggml whisper huggingface]'
May 08 12:19:18 localai local-ai[1333]: 12:19PM INF [llama-cpp] Attempting to load
May 08 12:19:18 localai local-ai[1333]: 12:19PM INF BackendLoader starting backend=llama-cpp modelID=qwen3-32b o.model=Qwen_Qwen3-32B-Q4_K_M.gguf
May 08 12:19:18 localai local-ai[1333]: 12:19PM INF [llama-cpp] attempting to load with CUDA variant
May 08 12:19:56 localai local-ai[1333]: 12:19PM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:43369: connect: connection refused""
May 08 12:19:58 localai local-ai[1333]: 12:19PM ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: grpc service not ready
May 08 12:20:37 localai local-ai[1333]: 12:20PM INF [llama-cpp] Loads OK
Same hardware works without issues with GPT4All.