-
Notifications
You must be signed in to change notification settings - Fork 1.2k
/
create_model_cards.py
87 lines (76 loc) · 3.56 KB
/
create_model_cards.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
import shutil
import os
import huggingface_hub
import pytest
import torch
from transformers import AutoModelForCausalLM
@pytest.mark.parametrize(
"model_name, base_model, dataset, training_logs",
[
(
"h2ogpt-oasst1-256-20b",
"EleutherAI/gpt-neox-20b",
"h2oai/openassistant_oasst1",
"https://huggingface.co/h2oai/h2ogpt-oasst1-256-20b/blob/main/gpt-neox-20b.openassistant_oasst1.json.1_epochs.5fc91911bc2bfaaf3b6c2de577c4b0ae45a07a4a.18.zip",
),
(
"h2ogpt-oasst1-512-12b",
"EleutherAI/pythia-12b",
"h2oai/openassistant_oasst1",
"https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/pythia-12b.openassistant_oasst1.json.1_epochs.d45a9d34d34534e076cc6797614b322bd0efb11c.15.zip",
),
(
"h2ogpt-oig-oasst1-256-12b",
"EleutherAI/pythia-12b-deduped",
"h2oai/h2ogpt-oig-oasst1-instruct-cleaned-v1",
"https://huggingface.co/h2oai/h2ogpt-oig-oasst1-256-12b/blob/main/pythia-12b-deduped.h2ogpt-oig-oasst1-instruct-cleaned-v1.json.1_epochs.5fc91911bc2bfaaf3b6c2de577c4b0ae45a07a4a.17.zip",
),
(
"h2ogpt-oig-oasst1-256-6.9b",
"EleutherAI/pythia-6.9b",
"h2oai/h2ogpt-oig-oasst1-instruct-cleaned-v1",
"https://huggingface.co/h2oai/h2ogpt-oig-oasst1-256-6.9b/blob/main/pythia-6.9b.h2ogpt-oig-oasst1-instruct-cleaned-v1.json.1_epochs.5fc91911bc2bfaaf3b6c2de577c4b0ae45a07a4a.9.zip",
),
],
)
def test_create_model_cards(model_name, base_model, dataset, training_logs):
model_size = model_name.split("-")[-1].upper()
assert "B" == model_size[-1]
assert int(model_size[-2]) >= 0
assert os.path.exists("README-template.md"), "must be running this test from the model dir."
shutil.rmtree(model_name, ignore_errors=True)
repo = huggingface_hub.Repository(
local_dir=model_name,
clone_from="h2oai/%s" % model_name,
skip_lfs_files=True,
token=True,
)
repo.git_pull()
model = AutoModelForCausalLM.from_pretrained("h2oai/%s" % model_name,
local_files_only=False,
torch_dtype=torch.float16,
device_map="auto")
model_arch = str(model)
model_config = str(model.config)
with open("README-template.md", "r") as f:
content = f.read()
assert "<<MODEL_NAME>>" in content
content = content.replace("<<MODEL_NAME>>", model_name)
assert "<<MODEL_SIZE>>" in content
content = content.replace("<<MODEL_SIZE>>", model_size[:-1])
assert "<<BASE_MODEL>>" in content
content = content.replace("<<BASE_MODEL>>", f"[{base_model}](https://huggingface.co/{base_model})")
assert "<<DATASET>>" in content
content = content.replace("<<DATASET>>", f"[{dataset}](https://huggingface.co/{dataset})")
assert "<<MODEL_ARCH>>" in content
content = content.replace("<<MODEL_ARCH>>", model_arch)
assert "<<MODEL_CONFIG>>" in content
content = content.replace("<<MODEL_CONFIG>>", model_config)
assert "<<TRAINING_LOGS>>" in content
content = content.replace("<<TRAINING_LOGS>>", training_logs)
assert "<<" not in content
assert ">>" not in content
with open(os.path.join(model_name, "README.md"), "w") as f:
f.write(content)
repo.commit("Update README.md")
repo.push_to_hub()