Skip to content

zjunlp/Knowledge2Data

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

👉 Knowledge2Data 👈

Spatial Knowledge Graph-Guided Multimodal Synthesis

Awesome License: MIT

ProjectPaperHuggingFaceOverviewQuickstartCitation

Table of Contents

🔔News

  • 2025-02-28, We release the paper.

🌟Overview

⏩Quickstart

Data

Get training data and test data from HuggingFace: https://huggingface.co/datasets/zjunlp/Knowledge2Data

Installation

git clone https://github.com/zjunlp/Knowledge2Data
cd Knowledge2Data
conda create -n skg python==3.9
conda activate skg
pip install -r requirements.txt

Download the models

Download the following models from HuggingFace

🎯 Model Name 🤗 HuggingFace
Diffusers-generation-text-box gligen/diffusers-generation-text-box
Sam-vit-base stabilityai/stable-diffusion-xl-refiner-1.0
Stable-diffusion-xl-refiner facebook/sam-vit-base

Export the environment variables.

cd src
export OPENAI_API_KEY="YOUR_API_KEY"
export SKG_HF_MODELS="LOCAL_HUGGINGFACE_MODELS_DIR"

Generate Spatial KG and multimodal synthetic data.

Execute script to generate Spatial KG.

sh run_skg.sh

You can also customize objects and their spatial relationships to form Spatial KG. Save the file format as a JSON file similar to "src/data/skg_demo.json".

Execute script to multimodal synthetic data.

sh run_data.sh

For custom data, only the input file parameters "--input_file" need to be modified.

You can find generated data in "src/data" and images in "src/img_generations" as default. If you want to generate more data, you can modify the parameters including "--num_scenes" (generate_scenes.py) and "--repeats" (generate_images.py).

🌻Acknowledgement

This project is based on open-source projects including LLM-groundedDiffusion. Thanks for their great contributions!

🚩Citation

Please cite the following paper if you use this project in your work.

@misc{xue2025spatialknowledgegraphguidedmultimodal,
      title={Spatial Knowledge Graph-Guided Multimodal Synthesis}, 
      author={Yida Xue and Zhen Bi and Jinnan Yang and Jungang Lou and Huajun Chen and Ningyu Zhang},
      year={2025},
      eprint={2505.22633},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.22633}, 
}

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •