You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
note down ControlNet related IDs, and "preprocessor" key specifically for those
this is an addition of a specialty non-altering optional key, which is considered a non-breaking change to the document (optional specialty keys are always allowed regardless of whether they are documented or not).
Copy file name to clipboardExpand all lines: README.md
+5-2Lines changed: 5 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,9 @@ The data within should identify clearly (A) the type and function of a model, cl
6
6
7
7
## Version
8
8
9
-
This is specification version 1.0.0.
9
+
This is specification version 1.0.1.
10
+
11
+
Versions are (approximately) SemVer: `(backward-breaking).(forward-breaking).(non-breaking-change)`.
10
12
11
13
## Technical Placement
12
14
@@ -85,6 +87,7 @@ This specification defines 3 categories of key: **MUST**, **SHOULD**, **CAN**
85
87
|`prediction_type`|**CAN**| In Stable Diffusion, `v` or `epsilon`. Other model classes may have their own concepts that apply. |`v`, `epsilon`|
86
88
|`timestep_range`|**CAN**| If a model is tuned on a sub-section of possible timesteps (Timestep-Expert Models), identify it here, in the format `<min>,<max>`. |`500,999`, `0,499`|
87
89
|`encoder_layer`|**CAN**| (Specialty) for "clip skip" in Stable Diffusion models, or similar practice in other models like it, this can be applied where relevant to identify that a non-standard layer of an encoder model should be used (so for example value `2` in an SD model indicates `clip_skip=2` should be used). |`2`|
90
+
|`preprocessor`|**CAN**| (Specialty) for "ControlNet" or similar model-adapter types that require preprocessing, this is an indicator of the preprocessing type, as a simple text identifier. Should not identify exact tool (eg "MiDaS"), just the broad type (eg "depth"). |`depth`, `canny`|
88
91
89
92
#### Text-Prediction Models
90
93
@@ -101,7 +104,7 @@ The following is a list of common Architecture ID values, both to serve as a ref
101
104
102
105
-**Stable Diffusion:**`stable-diffusion-v1`, or change `v1` to any of: `v1`, `v1-inpainting``v2-512`, `v2-768-v`, `v2-depth`, `v2-inpainting`, `v2-unclip-h`, `v2-unclip-l`, `xl-v1-base`, `xl-v1-refiner`
103
106
-**Stable Diffusion Components:**`stable-diffusion-xl-v1/vae` (change `stable-diffusion-xl-v1` to the base model architecture)
104
-
-**Stable Diffusion Adapters:**`stable-diffusion-v1/lora`, `stable-diffusion-v1/textual-inversion` (change `stable-diffusion-v1` to the base model architecture)
107
+
-**Stable Diffusion Adapters:**`stable-diffusion-v1/lora`, `stable-diffusion-v1/textual-inversion`, `stable-diffusion-v1/controlnet`, `stable-diffusion-v1/control-lora` (change `stable-diffusion-v1` to the base model architecture)
105
108
-**Language Models:**`gpt-neo-x`
106
109
107
110
(Note: relevant project leads for well-known model formats are welcomed to PR additions to this list)
0 commit comments