File tree Expand file tree Collapse file tree 3 files changed +5
-302
lines changed
vllm/model_executor/models Expand file tree Collapse file tree 3 files changed +5
-302
lines changed Original file line number Diff line number Diff line change 24
24
"GPTBigCodeForCausalLM" : ("gpt_bigcode" , "GPTBigCodeForCausalLM" ),
25
25
"GPTJForCausalLM" : ("gpt_j" , "GPTJForCausalLM" ),
26
26
"GPTNeoXForCausalLM" : ("gpt_neox" , "GPTNeoXForCausalLM" ),
27
- "InternLMForCausalLM" : ("internlm " , "InternLMForCausalLM " ),
27
+ "InternLMForCausalLM" : ("llama " , "LlamaForCausalLM " ),
28
28
"InternLM2ForCausalLM" : ("internlm2" , "InternLM2ForCausalLM" ),
29
29
"LlamaForCausalLM" : ("llama" , "LlamaForCausalLM" ),
30
30
# For decapoda-research/llama-*
Load Diff This file was deleted.
Original file line number Diff line number Diff line change @@ -91,6 +91,7 @@ def __init__(
91
91
rope_scaling : Optional [Dict [str , Any ]] = None ,
92
92
max_position_embeddings : int = 8192 ,
93
93
linear_method : Optional [LinearMethodBase ] = None ,
94
+ bias : bool = False ,
94
95
) -> None :
95
96
super ().__init__ ()
96
97
self .hidden_size = hidden_size
@@ -120,13 +121,13 @@ def __init__(
120
121
self .head_dim ,
121
122
self .total_num_heads ,
122
123
self .total_num_kv_heads ,
123
- bias = False ,
124
+ bias = bias ,
124
125
linear_method = linear_method ,
125
126
)
126
127
self .o_proj = RowParallelLinear (
127
128
self .total_num_heads * self .head_dim ,
128
129
hidden_size ,
129
- bias = False ,
130
+ bias = bias ,
130
131
linear_method = linear_method ,
131
132
)
132
133
@@ -179,6 +180,7 @@ def __init__(
179
180
rope_scaling = rope_scaling ,
180
181
max_position_embeddings = max_position_embeddings ,
181
182
linear_method = linear_method ,
183
+ bias = getattr (config , "bias" , False ),
182
184
)
183
185
self .mlp = LlamaMLP (
184
186
hidden_size = self .hidden_size ,
You can’t perform that action at this time.
0 commit comments