Skip to content

Clean install fails to run any model #5225

Open
@tescophil

Description

@tescophil

LocalAI version:

Latest version I stalled yesterday no idea how to get the tag/commit

Environment, CPU architecture, OS, and Version:
Linux desktop-garage 4.19.0-12-amd64 #1 SMP Debian 4.19.152-1 (2020-10-18) x86_64 GNU/Linux

Describe the bug
Clean install and l download of several models, all of them fail to load

To Reproduce
Install and attempt to run a model

Expected behavior
I expect a model to load

Logs
10:14AM INF Trying to load the model 'minicpm-v-2_6' with the backend '[llama-cpp llama-cpp-fallback piper silero-vad stablediffusion-ggml whisper bark-cpp huggingface]'
10:14AM INF [llama-cpp] Attempting to load
10:14AM INF BackendLoader starting backend=llama-cpp modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:14AM INF [llama-cpp] attempting to load with AVX variant
10:14AM INF Success ip=10.8.1.10 latency=767.980683ms method=POST status=200 url=/v1/chat/completions
10:14AM INF Success ip=10.8.1.10 latency="29.688µs" method=GET status=200 url=/static/favicon.svg
10:15AM INF Trying to load the model 'minicpm-v-2_6' with the backend '[llama-cpp llama-cpp-fallback whisper bark-cpp piper silero-vad stablediffusion-ggml huggingface]'
10:15AM INF [llama-cpp] Attempting to load
10:15AM INF BackendLoader starting backend=llama-cpp modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:15AM ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:15AM INF [llama-cpp] attempting to load with AVX variant
10:15AM ERR [llama-cpp] Failed loading model, trying with fallback 'llama-cpp-fallback', error: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:15AM INF [llama-cpp] Fails: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:15AM INF [llama-cpp-fallback] Attempting to load
10:15AM INF BackendLoader starting backend=llama-cpp-fallback modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:16AM INF [llama-cpp] Fails: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:16AM INF [llama-cpp-fallback] Attempting to load
10:16AM INF BackendLoader starting backend=llama-cpp-fallback modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:16AM INF [llama-cpp-fallback] Fails: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:16AM INF [piper] Attempting to load
10:16AM INF BackendLoader starting backend=piper modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:16AM INF [llama-cpp-fallback] Fails: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:16AM INF [whisper] Attempting to load
10:16AM INF BackendLoader starting backend=whisper modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:17AM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:36117: connect: connection refused""
10:17AM INF [piper] Fails: failed to load model with internal loader: grpc service not ready
10:17AM INF [silero-vad] Attempting to load
10:17AM INF BackendLoader starting backend=silero-vad modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:17AM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:42421: connect: connection refused""
10:17AM INF [whisper] Fails: failed to load model with internal loader: grpc service not ready
10:17AM INF [bark-cpp] Attempting to load
10:17AM INF BackendLoader starting backend=bark-cpp modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:18AM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:39127: connect: connection refused""
10:18AM INF [silero-vad] Fails: failed to load model with internal loader: grpc service not ready
10:18AM INF [stablediffusion-ggml] Attempting to load
10:18AM INF BackendLoader starting backend=stablediffusion-ggml modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:19AM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:45537: connect: connection refused""
10:19AM INF [bark-cpp] Fails: failed to load model with internal loader: grpc service not ready
10:19AM INF [piper] Attempting to load
10:19AM INF BackendLoader starting backend=piper modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:19AM INF [stablediffusion-ggml] Fails: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
10:19AM INF [whisper] Attempting to load
10:19AM INF BackendLoader starting backend=whisper modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf
10:19AM ERR failed starting/connecting to the gRPC service error="rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:39659: connect: connection refused""
10:19AM INF [piper] Fails: failed to load model with internal loader: grpc service not ready
10:19AM INF [silero-vad] Attempting to load
10:19AM INF BackendLoader starting backend=silero-vad modelID=minicpm-v-2_6 o.model=minicpm-v-2_6-Q4_K_M.gguf

Additional context
The basic documentation is very sparse and the default localai.env installed does not match up with the default master on GitHub. all the variables are prefixes with LOCALAI_ in the documentation, but the files installed by default does not e.g. LOCALAI_THREADS VS. THREADS. Also the installer runs the system as a service, but the docs don't mention this anywhere, rather it says run from the command line 'local-ai run' which fails as the service is already running.

I think this may be the same issue as #5216

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions