Skip to content

Please improve error messages and getting started documentation #1416

@chris-hatton

Description

@chris-hatton

Is your feature request related to a problem? Please describe.
Yes; I can't get your sample working. I'm using a cublas build with an Nvidia GPU; I've followed the setup carefully, and I don't see any log errors related to initialising the GPU; instead I always see failed to load model/EOF/bad magic. I've tried many models including luna-ai-llama2 from your sample, and have created the four files specified.

Describe the solution you'd like
Please Improve the error messages surfaced by Local.AI, and your documentation.
It looks like errors are heavily obfuscated at the moment due to the internal architecture of Local.AI You seen to be separating front and back end as two separate services, and the back-end does a poor job of surfacing error messages to the front end.

Describe alternatives you've considered
I have no alternatives, except to give up & go home.

Additional context
I'm not a complete dummy; I've had llama.cpp working on Metal and CPU before, but Local.AI's documentation and error messages leave a lot to be desired; I feel like I'm flying completely blind. Sorry to say your 'Getting started' documentation is not very well written and fails to establish vital facts for beginners such as:

  • What's the relationship between the naming of the model and the ID shown
  • Are the four files really vital? If so why doesn't Local.AI stop as soon as they're missing, or make a very clear error log about this
  • What's f16 mode? Do I have to enable it when working with a GPU?
  • What's the difference between using CUDA 11 or 12 build?

Metadata

Metadata

Labels

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions