Objaverse asset annotator and importer for use in THOR.
-
use
gpt-4o
-
set
get_best_synset
to False
Go to objathor/objathor/main.py
, around line 100, add the argument to the function get_initial_annotation
anno, urls = get_initial_annotation(
uid,
thumbnail_urls_cfg=dict(
base_url=render_dir,
view_indices=[str(float(angle)) for angle in render_angles],
local_renders=True,
),
>>> get_best_synset=False
)
Install ai2thor:
pip install --extra-index-url https://ai2thor-pypi.allenai.org ai2thor==0+40679c517859e09c1f2a5e39b65ee7f33fcfdd48
Install other dependencies:
pip install objathor[annotation]
Here the following extras are installed: annotation
to use openai to generate annotations. Also for annotation functionality you must install nltk
Install nltk. To generate renders and convert 'glb' models in the conversion pipeline you must Install Blender .
From source:
pip install -e ".[annotation]"
We recommend setting an environment variable with your OpenAI key:
export OPENAI_API_KEY=[Your key]
If you're planning to annotate objects for which we don't have pre-generated views in S3 (see an example under Annotation below), we also need to install blender either as an extra (as shown above) or as an application:
Installing the Blender
as a module:
pip install bpy
Installing the Blender as a module, requires a python 3.10
environment.
Or installing blender as an application: Blender install instructions
If application is not in the cannonical directories you may need to pass blender_installation_path
to scripts that use Blender.
Install nltk
on this commit by running:
pip install git+https://github.com/nltk/nltk@582e6e35f0e6c984b44ec49dcb8846d9c011d0a8
During the first run, NLTK dependencies are automatically installed, but we can also install them ahead:
python -c "import nltk; nltk.download('punkt'); nltk.download('wordnet2022'); nltk.download('brown'); nltk.download('averaged_perceptron_tagger')"
For automatic annotation to assign likely synsets given the automatically generated asset description, we can pre-install pre-generated embeddings for all synset definitions (this can be useful if we cannot write into our home directory at run time):
mkdir ~/.objathor_data
curl https://prior-datasets.s3.us-east-2.amazonaws.com/vida-synset-embeddings/synset_definition_embeddings_single.pkl.gz -o ~/.objathor_data/synset_definition_embeddings_single.pkl.gz
Assuming we're running on a remote Linux server, we can pre-download the THOR binaries with:
python -c "from ai2thor.controller import Controller; from objathor.constants import THOR_COMMIT_ID; c=Controller(download_only=True, platform='CloudRendering', commit_id=THOR_COMMIT_ID)"
(platform='OSXIntel64'
would be used for a MacOS environment).
You must install the annotation
extra requirement through pip, ad have blender installed,
either standalone or as a module. The following command will generate annotation, via GPT-4,
and also generate the conversion to a valid THOR asset.
OUTPUT_DIR=/path/to/output
python -m objathor.main \
--uid 0070ac4bf50b496387365843d4bf5432 \
--output "$OUTPUT_DIR"
From the repository root run:
python
-m objathor.asset_conversion.pipeline_to_thor
--object_ids=000074a334c541878360457c672b6c2e
--output_dir=<some_absolute_path>
--extension=.msgpack.gz
--annotations=<annotations_file_path>
--live
--blender_as_module
Where object_ids
is a string of comma separated list of Objaverse
object ids to process.
output_dir
is an absolute path indicating where to write the output of the conversion.
annotations
is optional, and is the path to an annotations file as generated by the process described above.
Run python -m objathor.asset_conversion.pipeline_to_thor --help
for other options.