Skip to content

Commit 4efbac6

Browse files
authored
Migrate AquilaForCausalLM to LlamaForCausalLM (#2867)
1 parent 87069cc commit 4efbac6

File tree

5 files changed

+2
-417
lines changed

5 files changed

+2
-417
lines changed

vllm/model_executor/models/__init__.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@
1010

1111
# Architecture -> (module, class).
1212
_MODELS = {
13-
"AquilaModel": ("aquila", "AquilaForCausalLM"),
14-
"AquilaForCausalLM": ("aquila", "AquilaForCausalLM"), # AquilaChat2
13+
"AquilaModel": ("llama", "LlamaForCausalLM"),
14+
"AquilaForCausalLM": ("llama", "LlamaForCausalLM"), # AquilaChat2
1515
"BaiChuanForCausalLM": ("baichuan", "BaiChuanForCausalLM"), # baichuan-7b
1616
"BaichuanForCausalLM": ("baichuan", "BaichuanForCausalLM"), # baichuan-13b
1717
"BloomForCausalLM": ("bloom", "BloomForCausalLM"),
@@ -41,7 +41,6 @@
4141
"Qwen2ForCausalLM": ("qwen2", "Qwen2ForCausalLM"),
4242
"RWForCausalLM": ("falcon", "FalconForCausalLM"),
4343
"StableLMEpochForCausalLM": ("stablelm", "StablelmForCausalLM"),
44-
"YiForCausalLM": ("yi", "YiForCausalLM")
4544
}
4645

4746
# Models not supported by ROCm.

vllm/model_executor/models/aquila.py

Lines changed: 0 additions & 342 deletions
This file was deleted.

vllm/transformers_utils/config.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55
from vllm.transformers_utils.configs import *
66

77
_CONFIG_REGISTRY = {
8-
"aquila": AquilaConfig,
98
"baichuan": BaiChuanConfig,
109
"chatglm": ChatGLMConfig,
1110
"mpt": MPTConfig,

vllm/transformers_utils/configs/__init__.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
from vllm.transformers_utils.configs.aquila import AquilaConfig
21
from vllm.transformers_utils.configs.baichuan import BaiChuanConfig
32
from vllm.transformers_utils.configs.chatglm import ChatGLMConfig
43
from vllm.transformers_utils.configs.mpt import MPTConfig
@@ -9,7 +8,6 @@
98
from vllm.transformers_utils.configs.falcon import RWConfig
109

1110
__all__ = [
12-
"AquilaConfig",
1311
"BaiChuanConfig",
1412
"ChatGLMConfig",
1513
"MPTConfig",

0 commit comments

Comments
 (0)