Skip to content

Commit 1de713c

Browse files
committed
Fix errors when use verl to train GLM4.1v model
* Support glm4v load from AutoModelForVision2Seq * Set glm4v model _checkpoint_conversion_mapping attr from None to {}
1 parent bbca978 commit 1de713c

File tree

3 files changed

+5
-4
lines changed

3 files changed

+5
-4
lines changed

src/transformers/models/auto/modeling_auto.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -877,6 +877,7 @@
877877
("blip-2", "Blip2ForConditionalGeneration"),
878878
("chameleon", "ChameleonForConditionalGeneration"),
879879
("git", "GitForCausalLM"),
880+
("glm4v", "Glm4vForConditionalGeneration"),
880881
("idefics2", "Idefics2ForConditionalGeneration"),
881882
("idefics3", "Idefics3ForConditionalGeneration"),
882883
("instructblip", "InstructBlipForConditionalGeneration"),

src/transformers/models/glm4v/modeling_glm4v.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -949,7 +949,7 @@ def forward(
949949
@auto_docstring
950950
class Glm4vModel(Glm4vPreTrainedModel):
951951
base_model_prefix = ""
952-
_checkpoint_conversion_mapping = None
952+
_checkpoint_conversion_mapping = {}
953953
config_class = Glm4vConfig
954954
_no_split_modules = ["Glm4vTextDecoderLayer", "Glm4vVisionBlock"]
955955

@@ -1382,7 +1382,7 @@ class Glm4vCausalLMOutputWithPast(ModelOutput):
13821382

13831383

13841384
class Glm4vForConditionalGeneration(Glm4vPreTrainedModel, GenerationMixin):
1385-
_checkpoint_conversion_mapping = None
1385+
_checkpoint_conversion_mapping = {}
13861386
_tied_weights_keys = ["lm_head.weight"]
13871387

13881388
def __init__(self, config):

src/transformers/models/glm4v/modular_glm4v.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1001,7 +1001,7 @@ def forward(
10011001

10021002

10031003
class Glm4vModel(Qwen2_5_VLModel):
1004-
_checkpoint_conversion_mapping = None
1004+
_checkpoint_conversion_mapping = {}
10051005
_no_split_modules = ["Glm4vTextDecoderLayer", "Glm4vVisionBlock"]
10061006

10071007
def __init__(self, config):
@@ -1356,7 +1356,7 @@ class Glm4vCausalLMOutputWithPast(Qwen2_5_VLCausalLMOutputWithPast):
13561356

13571357

13581358
class Glm4vForConditionalGeneration(Qwen2_5_VLForConditionalGeneration):
1359-
_checkpoint_conversion_mapping = None
1359+
_checkpoint_conversion_mapping = {}
13601360

13611361
def forward(
13621362
self,

0 commit comments

Comments
 (0)