Description
I accidentally closed the previous issue.
I want to automate this I2V model into my workflow, so when I try to run Tora's I2V model example locally on an RTX 4090, I always encounter CUDA out of memory issues. Can I modify some parameters in Tora/sat/sample_video.py or Tora/sat/configs/tora/inference_sparse.yaml or Tora/sat/configs/tora/model/cogvideox_5b_tora_i2v.yaml to make this model run on an RTX 4090 with 24GB VRAM?
Although I am aware that the tora I2V model can successfully generate corresponding videos in the comfyui-tora, I still want to run it on my local machine through Python code to better integrate this model into my workflow.
You could try the diffusers-version with "--enable_model_cpu_offload" option. https://github.com/alibaba/Tora/blob/main/diffusers-version/README.md
If still OOM, try "--enable_sequential_cpu_offload" instead.
Originally posted by @Leojc in #29
According to Leojc's reply, I checked the Tora diffusers-version. However, the diffusers-version only supports T2V and not I2V? Or did I miss something?