Skip to content

Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.

License

Notifications You must be signed in to change notification settings

tjennings/Coreco_LLaVA

 
 

Repository files navigation

Coreco's LLaVA captioning fork

This is a total hack to use LLaVA to caption large image datasets To use it you must:

  • Follow the LLaVA installation instructions to set up this code base, acquire the llama model, and merge their delta.
  • Run the following command:
python3 -m llava.eval.run_llava --path /path/to/images --model_path /path/to/model

Full options:

usage: described [-h] --path PATH [--overwrite] [--model_name MODEL_NAME] [--prefix PREFIX] [--suffix SUFFIX] [--resize RESIZE]

options:
  -h, --help            show this help message and exit
  --path PATH           Path to images to be captioned
  --overwrite           Overwrite existing captions
  --model_name MODEL_NAME
                        path to the model
  --prefix PREFIX       a string applied at the beginning of each caption
  --suffix SUFFIX       a string applied at the end of each caption
  --resize RESIZE       additionally, resize and save the image where the longest side is the provided maximum

About

Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 93.0%
  • JavaScript 2.7%
  • HTML 2.1%
  • Shell 1.7%
  • CSS 0.5%