This repository packages NVIDIA's ChronoEdit-14B Diffusers image editing model for deployment on Replicate using Cog. ChronoEdit treats image editing as a temporal generation task to preserve structure while applying rich edits, and ships here with NVIDIA's 8-step distillation LoRA for fast inference.
- Loads the official ChronoEdit diffusers pipeline (
nvidia/ChronoEdit-14B-Diffusers). - Applies the provided 8-step distillation LoRA for accelerated inference.
- Optional prompt expansion via
Qwen/Qwen3-VL-8B-Instruct(matches the NVIDIA Gradio demo). - Caches model weights under
checkpoints/to avoid repeated downloads during local testing.
- Docker (required by Cog)
- An NVIDIA GPU with at least 40 GB of VRAM is recommended. Smaller GPUs may OOM.
- Python is not required directly; Cog will manage the runtime inside Docker.
Install Cog if you have not already:
curl -o /usr/local/bin/cog -L https://github.com/replicate/cog/releases/latest/download/cog_$(uname -s)_$(uname -m)
chmod +x /usr/local/bin/cogThe first cog predict run downloads ~20 GB of model assets. To warm the cache ahead of time:
cog predict -i image=@path/to/example.png -i prompt="Describe the edit you want"Weights will be stored in checkpoints/ChronoEdit-14B-Diffusers/ so subsequent runs reuse them.
cog predict \
-i image=@input.png \
-i prompt="Replace the flowers with glass sculptures" \
-i enable_prompt_expansion=true \
-i num_steps=8Outputs are written to output.jpg by default. Set enable_prompt_expansion=false to skip the Qwen-based rewrite.
predict.py– Cog predictor loading ChronoEdit, LoRA weights, and optional prompt enhancer.chronoedit_diffusers/– Vendorized ChronoEdit pipeline modules referenced by the predictor.requirements.txt– Python dependencies installed inside the Cog container.cog.yaml– Runtime configuration (Python 3.11, CUDA 12.4 base image, GPU enabled).
After testing locally:
cog login
cog push r8.im/<username>/chronoeditReplace <username> with your Replicate handle. Follow Replicate's CLI prompts to publish the model card.
- The base image uses CUDA 12.4; adjust
cog.yamlif your target environment requires a different CUDA version. - Prompt expansion loads a large vision-language model; disable it to reduce memory usage or cold-start time.
- Grande-scale edits remain computationally intensive despite distillation—expect multi-minute cold starts.
For more details on ChronoEdit itself, see NVIDIA's ChronoEdit model card and project page.