Releases: modelscope/ms-swift
Releases · modelscope/ms-swift
v2.4.1
English Version
New Features:
- Inference and deployment support for logprobs.
- RLHF support for lazy_tokenize.
- Multimodal model support for neftune.
- dynamic_eos compatibility with glm4 series and other models.
New Models:
- mplug-owl3, best practices can be found here.
- yi-coder 1.5b, base/chat model of 9b.
- minicpm3-4b.
- reflection-llama3.1-70b.
中文版
新功能:
- 推理和部署支持 logprobs。
- RLHF支持lazy_tokenize。
- 多模态模型支持neftune。
- dynamic_eos兼容glm4系列等模型。
新模型:
- mplug-owl3,最佳实践可以查看这里。
- yi-coder 1.5b、9b 的base/chat模型。
- minicpm3-4b。
- reflection-llama3.1-70b。
What's Changed
- Fix push_to_hub when last-checkpoint by @tastelikefeet in #1897
- support custom quantized dataset by @tastelikefeet in #1893
- fix push_to_ms by @tastelikefeet in #1901
- support logprobs by @Jintao-Huang in #1900
- deepspeed use cosine lr_schduler by @Jintao-Huang in #1907
- update docs by @Jintao-Huang in #1908
- fix web-ui push to hub strategy by @tastelikefeet in #1909
- Refactor docs by @tastelikefeet in #1912
- refactor docs by @tastelikefeet in #1915
- [TorchAcc] perf: use xm.save instead of torch.save by @baoleai in #1916
- update wechat by @tastelikefeet in #1925
- update docs & fix bug by @Jintao-Huang in #1926
- [TorchAcc] fix: fix the judegement of fsdp_num by @baoleai in #1903
- Support deploy & logprobs by @Jintao-Huang in #1833
- fix typing by @Jintao-Huang in #1933
- fix swift deploy by @Jintao-Huang in #1936
- update yi-coder by @Jintao-Huang in #1942
- fix lmdeploy seed by @Jintao-Huang in #1945
- fix do_sample by @Jintao-Huang in #1946
- refactor rlhf by @hjh0119 in #1885
- fix file rename error in megatron when there are multi process by @Zhikaiiii in #1948
- fix qwen2-vl & video by @Jintao-Huang in #1950
- support dynamic_eos by @Jintao-Huang in #1947
- fix rlhf by @hjh0119 in #1949
- Support minicpm 3 by @Jintao-Huang in #1952
- Add lazy_tokenize to RLHF by @tastelikefeet in #1956
- Fix data info print in rlhf by @tastelikefeet in #1964
- Fix the lora hook by @tastelikefeet in #1963
- fix bugs by @Jintao-Huang in #1959
- support mplug_owl3 by @Jintao-Huang in #1957
- update docs by @Jintao-Huang in #1970
- Add reflection model by @tastelikefeet in #1973
- fix typo by @Jintao-Huang in #1980
Full Changelog: v2.4.0...v2.4.1
v2.4.0
English Version
New Features:
- Support for Liger, which accommodates models like LLaMA, Qwen, Mistral, etc., and reduces memory usage by 10% to 60%.
- Support for custom loss function training using a registration mechanism.
- Training now supports pushing models to ModelScope and HuggingFace.
- Support for the
freeze_vit
parameter to control the behavior of full parameter training for multimodal models.
New Models:
- Qwen2-VL series includes GPTQ/AWQ quantized models. For best practices, see here.
- InternVL2 AWQ quantized models.
New Datasets:
- qwen2-pro series
中文版
新特性:
- 支持 Liger训练LLaMA、Qwen、Mistral 等模型,内存使用降低 10% 至 60%。
- 支持使用注册机制进行自定义损失函数的训练。
- 训练支持将模型推送至 ModelScope 和 HuggingFace。
- 支持 freeze_vit 参数,以控制多模态模型全参数训练的行为。
新模型:
- Qwen2-VL 系列包括 GPTQ/AWQ 量化模型,最佳实践可以查看这里。
- InternVL2 AWQ 量化模型。
新数据集:
- qwen2-pro 系列
What's Changed
- compat with vllm==0.5.5 by @Jintao-Huang in #1812
- Support zero2 offload by @Jintao-Huang in #1814
- fix mp+ddp & resume_from_checkpoint by @Jintao-Huang in #1815
- fix preprocess_num_proc by @Jintao-Huang in #1818
- Support liger by @tastelikefeet in #1819
- fix dora deployment by @tastelikefeet in #1821
- Support register loss func by @Jintao-Huang in #1822
- use default-lora by @Jintao-Huang in #1823
- fix minicpm-v 2.6 infer device_map by @Jintao-Huang in #1832
- Fix code by @tastelikefeet in #1824
- fix inject by @tastelikefeet in #1835
- support qwen2-pro dataset by @Jintao-Huang in #1834
- add ddp_timeout parameter by @tastelikefeet in #1836
- fix internlm-xcomposer rlhf by @hjh0119 in #1838
- Support eval_nproc by @tastelikefeet in #1843
- support qwen2-vl by @Jintao-Huang in #1842
- Add internvl2 awq models by @tastelikefeet in #1846
- Fix some datasets for streaming by @tastelikefeet in #1848
- Fix Pissa and OLoRA by @tastelikefeet in #1852
- Support qwen2 vl grounding by @tastelikefeet in #1854
- support qwen2-vl & video finetune by @Jintao-Huang in #1849
- Update new datasets by @tastelikefeet in #1855
- update qwen2-vl docs by @Jintao-Huang in #1856
- update qwen2-vl docs by @Jintao-Huang in #1858
- fix qwen2-vl docs by @Jintao-Huang in #1861
- fix requirements by @Jintao-Huang in #1864
- update docs qwen2-vl by @Jintao-Huang in #1869
- Support faster data map by @tastelikefeet in #1871
- [TorchAcc] fix serveral bugs for torchacc FSDP. by @baoleai in #1872
- Add train record by @tastelikefeet in #1873
- Fix num_proc by @Jintao-Huang in #1874
- Fix neftune doc by @tastelikefeet in #1875
- add duet by @tastelikefeet in #1877
- use model.generation_config by @Jintao-Huang in #1850
- Support freeze vit by @Jintao-Huang in #1880
- support qwen2-vl gptq awq by @Jintao-Huang in #1884
- Refactor push_to_hub by @tastelikefeet in #1883
- Fix push to hub logic by @tastelikefeet in #1888
- add vllm lmdeploy benchmark by @Jintao-Huang in #1889
- Add some warnings and fix RLHF by @tastelikefeet in #1890
Full Changelog: v2.3.2...v2.4.0
v2.3.2
English Version
New Features:
- ReFT support: achieves parameter efficiency that is 15× to 65× greater than LoRA.
- Multimodal model supports zero3.
- Supports using environment variables to control parameters such as hd_num, max_num, and video_segments.
New Models:
- longwriter-glm4-9b, longwriter-llama3_1-8b
- phi3_5-mini-instruct, phi3_5-moe-instruct, phi3_5-vision-instruct
- llava-onevision-qwen2-0_5b-ov, llava-onevision-qwen2-7b-ov, llava-onevision-qwen2-72b-ov
New Datasets:
- longwriter-6k
- rlaif-v
- latex-ocr-print, latex-ocr-handwrite
中文版
新功能:
- 支持ReFT,实现了比 LoRA 高 15 倍到 65 倍的参数效率。
- 多模态模型支持 zero3。
- 支持使用环境变量控制模型特有的参数,如 hd_num、max_num 和 video_segments。
新模型:
- longwriter-glm4-9b, longwriter-llama3_1-8b
- phi3_5-mini-instruct, phi3_5-moe-instruct, phi3_5-vision-instruct
- llava-onevision-qwen2-0_5b-ov, llava-onevision-qwen2-7b-ov, llava-onevision-qwen2-72b-ov
新数据集:
- longwriter-6k
- rlaif-v
- latex-ocr-print, latex-ocr-handwrite
What's Changed
- fix imports by @tastelikefeet in #1748
- compat with torch=1.12/1.13 by @Jintao-Huang in #1752
- update rlaif-v hf dataset by @hjh0119 in #1755
- fix lmdeploy: AssertionError: failed to match chat template, please explicit set chat_template_config by @Jintao-Huang in #1759
- use eager -> sdpa by @Jintao-Huang in #1764
- Fix GLM4 agent toolcall by @tastelikefeet in #1767
- Support LongWriter-llama3.1-8b and LongWriter-glm4-9b. by @DaozeZhang in #1762
- Support llava onevision by @Jintao-Huang in #1761
- [TorchAcc] fix: fix saving and loading checkpoint for full sft FSDP by @baoleai in #1765
- Fix deepseek-coder-v2-lite template by @Jintao-Huang in #1771
- Fix qwen2-audio & zero3 by @Jintao-Huang in #1774
- Fix zero3 & minicpm-v/internvl2/xcomposer by @Jintao-Huang in #1772
- fix infer dataset_test_ratio by @Jintao-Huang in #1779
- fix moe & gradient_checkpointing by @Jintao-Huang in #1782
- support phi3.5-vision by @Jintao-Huang in #1780
- ReFT by @tastelikefeet in #1785
- update doc by @tastelikefeet in #1789
- support qwen-vl & base64 by @Jintao-Huang in #1790
- fix yi-vl template by @Jintao-Huang in #1793
- fix bugs by @Jintao-Huang in #1794
- fix imports by @tastelikefeet in #1796
- fix history_roles by @Jintao-Huang in #1798
- fix mllm rlhf with full sft type by @hjh0119 in #1800
- fix CI by @tastelikefeet in #1797
- fix megatron_patch_path by @wning13 in #1804
- Support hd num by @Jintao-Huang in #1801
- Support Latex OCR dataset by @Jintao-Huang in #1810
- fix offline export by @wning13 in #1805
- fix by @tastelikefeet in #1811
New Contributors
Full Changelog: v2.3.1...v2.3.2
v2.3.1
English Version
New Features:
- ms-swift paper published: https://arxiv.org/abs/2408.05517
- Web-UI supports audio and video.
- Support for deploying audio and video models using the OpenAI API.
- Utilizes a new multimodal training framework.
- supports inference acceleration for video models (lmdeploy & internvl2 series).
New Models:
- idefics3-8b-llama3
- llava-hf 72b, 110b, llama3-llava
- deepseek-coder-v2, deepseek-coder-lite-v2, deepseek-v2
中文版
新功能:
- 发布了 ms-swift 论文:https://arxiv.org/abs/2408.05517
- Web-UI 支持音频和视频。
- 支持使用 OpenAI API 部署音频和视频模型。
- 采用新的多模态训练框架。
- 支持视频模型的推理加速(lmdeploy 和 internvl2 系列)。
新模型:
- idefics3-8b-llama3
- llava-hf 72b、110b、llama3-llava
- deepseek-coder-v2、deepseek-coder-lite-v2、deepseek-v2
What's Changed
- support video deploy by @Jintao-Huang in #1632
- fix infer deploy by @Jintao-Huang in #1665
- Fix bugs by @tastelikefeet in #1664
- Add generation config support for swift eval by @Yunnglin in #1659
- Add a bunch of small features by @tastelikefeet in #1669
- Support lmdeploy&vllm support video by @Jintao-Huang in #1668
- [TorchAcc] fix: add mark_step when GA is greater than 1 by @baoleai in #1677
- Support audio and video in web-ui by @tastelikefeet in #1681
- Fix lmdeploy deploy seed by @Jintao-Huang in #1678
- add paper link by @tastelikefeet in #1684
- set default writer_batch_size argument in mllm rlhf by @hjh0119 in #1689
- New multimodal framework by @Jintao-Huang in #1594
- fix video component by @tastelikefeet in #1692
- refactor internvl2 by @Jintao-Huang in #1625
- fix model dtype in rlhf by @hjh0119 in #1696
- Fix minicpmv 2.6 & video/plain-text bug by @Jintao-Huang in #1698
- Fix internvl2 device map by @Jintao-Huang in #1701
- Support deepseek-v2 base by @Jintao-Huang in #1708
- [TorchAcc] update data bucketing strategy by @baoleai in #1706
- [TorchAcc] consolidate FSDP ckpt after saving instead of during loading by @baoleai in #1709
- fix deploy stream by @Jintao-Huang in #1713
- update docs by @Jintao-Huang in #1714
- Fix mllm rlhf by @hjh0119 in #1707
- support idefics3 by @Jintao-Huang in #1720
- Fix minicpm-v-v2.6 device_map='auto' & agent deploy by @Jintao-Huang in #1733
- Fix kto by @hjh0119 in #1737
- support llava-hf 72b, 110b, llama3-llava-hf by @Jintao-Huang in #1741
- Fix glm4v rlhf by @hjh0119 in #1745
New Contributors
Full Changelog: v2.3.0...v2.3.1
v2.3.0
English Version
New Features
- Support for readthedocs documentation site at: https://swift.readthedocs.io/en/latest
- Support Megatron architecture training for QianWen series models, and added new
pt
command for pretraining. See docs: https://swift.readthedocs.io/en/latest/LLM/Megatron-training.html - Support LMDeploy for inference and deployment, improving inference acceleration for multi-modal models. See: https://swift.readthedocs.io/en/latest/Multi-Modal/LmDeploy-inference-acceleration.html
- Support passing lora target modules via regular expressions
- Support configuring max_memory usage for each GPU in device_map
export
command supports BitsAndBytes quantizationexport
command supports Ollama export: https://swift.readthedocs.io/en/latest/LLM/OLLaMA-Export.html- Support Q-GaLore algorithm
- Support RLHF training for multi-modal models: https://swift.readthedocs.io/en/latest/Multi-Modal/human-preference-alignment-training-documentation.html
- Support evaluation on 100+ datasets for multi-modal models: https://swift.readthedocs.io/en/latest/LLM/LLM-eval.html
- Support resizing input images when memory usage is too high for multi-modal models
- Modified default lora injection for multi-modal model training. Now takes effect on LLM and projector, results are better without significantly increasing training memory.
- Support PEFT 0.12, and added new tuner: fourierft
- Support rope-scaling for multi-modal models
- Support streaming processing of datasets to reduce memory usage, enable with
--streaming
- Support vLLM multi-modal inference and deployment
- Support grounding task for popular multi-modal models.
New Models
- qwen2-audio series
- qwen2-math
- codegeex4
- internvl2 series
- llava video
- xcomposer2.5
- cogvlm2-video
- numina-math
- mistral-nemo
- llama3.1 series
- mistral-large
- gemma-2-2b
- internlm2.5 1.8b 20b
- minicpm-v-v2_6-chat
Check: https://swift.readthedocs.io/en/latest/LLM/Supported-models-datasets.html
New Datasets
- zhihu-kol and zhihu-kol-filtered
- SA1B series multi-modal zh datasets
Check: https://swift.readthedocs.io/en/latest/LLM/Supported-models-datasets.html
中文版本
新功能
- 支持readthedocs文档库, 地址:https://swift.readthedocs.io/zh-cn/latest
- 支持千问系列模型的Megatron结构训练,并支持了新的pt命令用于预训练,详见文档:https://swift.readthedocs.io/zh-cn/latest/LLM/Megatron%E8%AE%AD%E7%BB%83%E6%96%87%E6%A1%A3.html
- 支持LMDeploy的推理和部署,更好地支持了多模态模型的推理加速,详见:https://swift.readthedocs.io/zh-cn/latest/Multi-Modal/LmDeploy%E6%8E%A8%E7%90%86%E5%8A%A0%E9%80%9F%E6%96%87%E6%A1%A3.html
- 支持以正则表达式方式传入lora target模块
- 支持配置device_map各GPU用量的max_memory
- export命令支持BitsAndBytes量化
- export命令支持Ollama导出:https://swift.readthedocs.io/zh-cn/latest/LLM/OLLAMA%E5%AF%BC%E5%87%BA%E6%96%87%E6%A1%A3.html
- 支持Q-GaLore算法
- 支持多模态模型的RLHF训练:https://swift.readthedocs.io/zh-cn/latest/Multi-Modal/%E4%BA%BA%E7%B1%BB%E5%81%8F%E5%A5%BD%E5%AF%B9%E9%BD%90%E8%AE%AD%E7%BB%83%E6%96%87%E6%A1%A3.html
- 支持多模态模型100+数据集的评测能力:https://swift.readthedocs.io/zh-cn/latest/LLM/LLM%E8%AF%84%E6%B5%8B%E6%96%87%E6%A1%A3.html
- 支持多模态模型显存占用过高时对输入图片进行缩放
- 修改了多模态模型训练的默认lora注入,目前对LLM和projector生效,不显著提高训练显存情况下效果更好
- 支持PEFT0.12,并支持了新的tuner:fourierft
- 支持多模态模型的rope-scaling
- 支持数据集的流式处理,降低显存消耗,使用--streaming开启
- 支持了vLLM的多模态推理部署能力
- 对部分多模态模型支持了grounding任务
新模型
- qwen2-audio系列模型
- qwen2-math
- codegeex4
- internvl2系列模型
- llava video
- xcomposer2.5
- cogvlm2-video
- numina-math
- mistral-nemo
- llama3.1系列
- mistral-large
- gemma-2-2b
- internlm2.5 1.8b 20b
- minicpm-v-v2_6-chat
新数据集
- zhihu-kol和zhihu-kol-filtered数据集
- SA1B系列中文多模态数据集
What's Changed
- fix dependency by @tastelikefeet in #1306
- support codegeex4 by @Jintao-Huang in #1305
- support internvl2 by @hjh0119 in #1304
- support llava video by @Jintao-Huang in #1307
- fix docs by @Jintao-Huang in #1309
- support lr_scheduler_kwargs by @Jintao-Huang in #1310
- Fix internvl2 template by @hjh0119 in #1308
- Fix bugs by @tastelikefeet in #1311
- support warmup_stable_decay by @Jintao-Huang in #1312
- Support xcomposer2.5 by @Jintao-Huang in #1287
- Fix bugs by @tastelikefeet in #1319
- fix bug by @tastelikefeet in #1320
- fix template by @Jintao-Huang in #1321
- support cogvlm2-video by @Jintao-Huang in #1318
- Fix bugs by @tastelikefeet in #1325
- fix web-ui by @tastelikefeet in #1327
- compatible with trl 0.9.6 by @hjh0119 in #1326
- compat with vllm==0.5.1 by @Jintao-Huang in #1329
- Update qrcode by @tastelikefeet in #1332
- fix florence model by @hjh0119 in #1334
- Relaxing requirements for trl by @hjh0119 in #1342
- fix xcomposer2.5 device_map by @Jintao-Huang in #1343
- support generation_info by @Jintao-Huang in #1344
- fix requirements by @Jintao-Huang in #1347
- readthedocs by @tastelikefeet in #1345
- fix sequence parallel get labels by @tastelikefeet in #1352
- fix filelock by @tastelikefeet in #1354
- Add pt command by @tastelikefeet in #1356
- fix generation_info efficiency by @Jintao-Huang in #1359
- fix sh ddp_backend by @Jintao-Huang in #1360
- support LLM & lmdeploy by @Jintao-Huang in #1272
- fix a file path by @tastelikefeet in #1363
- Internvl2 support video by @tastelikefeet in #1366
- fix openai api by @Jintao-Huang in #1367
- fix internvl2-40b by @hjh0119 in #1369
- fix vlm deploy lora & agent by @Jintao-Huang in #1371
- Support lora regex by @tastelikefeet in #1375
- Fix docs by @Jintao-Huang in #1380
- Fix FSDP; Add training percentage to jsonl logging; Add a web-ui component by @tastelikefeet in #1381
- Support max memory args by @tastelikefeet in #1382
- fix max_memory by @tastelikefeet in #1383
- Fix gpu assert calculation by @tastelikefeet in #1384
- fix dataset_sample & deploy stop_words by @Jintao-Huang in #1385
- fix internvl doc by @hjh0119 in #1394
- Fix link by @Jintao-Huang in #1397
- fix vllm==0.5.1 by @Jintao-Huang in #1404
- [TorchAcc] update accelerate API and add llama3-70B by @baoleai in #1400
- Support Ollama and BNB for export by @tastelikefeet in #1407
- Fix glm4v merge lora by @Jintao-Huang in #1410
- [TorchAcc] fix model download when using TorchAcc distributed training by @baoleai in #1408
- Support padding left by @tastelikefeet in #1414
- Fix ollama export by @tastelikefeet in #1416
- fix web-ui params by @tastelikefeet in #1417
- fix hub_token by @Jintao-Huang in #1420
- Update ms hub token by @Jintao-Huang in #1424
- Add numina math model by @tastelikefeet in #1421
- fix internvl template by @Jintao-Huang in #1433
- Internvl series models update by @hjh0119 in #1426
- fix internvl2 template by @Jintao-Huang in #1436
- Fix bug and make lazydataset more stable by @tastelikefeet in #1438
- Fix llava-hf by @tastelikefeet in #1439
- [WIP]Support Q-Galore by @tastelikefeet in #1440
-
- support deepspeed on ui 2. add tools to client_utils by @tastelikefeet in #1446
- fix r...
v2.2.5
New Features:
- Support for RLHF with multimodal models, including DPO, ORPO, SIMPO, and CPO
- SWIFT infer, SWIFT deploy support using lmdeploy for inference acceleration.
- Support the use of Megatron for performing PT and SFT on the Qwen2 series models.
- Support the grounding task for InternVL2/Qwen-VL-Chat models
New Models:
- mistral-nemo series, mistral-large
- llama3.1 series
New Datasets:
- sa1b-dense-caption, sa1b-paired-caption
- rlaif-v
- zhihu-kol, zhihu-kol-filtered
What's Changed
- Support api key by @Jintao-Huang in #1452
- Support mistral nemo series models by @tastelikefeet in #1454
- fix minicpm-v2.5 lora_target_modules by @Jintao-Huang in #1455
- Add two datasets by @tastelikefeet in #1459
- Update trl dependency version by @hjh0119 in #1463
- fix bugs by @Jintao-Huang in #1464
- fix yi1.5 by @Jintao-Huang in #1465
- Fix yi1.5 by @Jintao-Huang in #1467
- add activate and deactivate for part tuner by @tastelikefeet in #1470
- support llama3.1 by @Jintao-Huang in #1475
- support megatron by @Jintao-Huang in #1365
- fix megatron by @Jintao-Huang in #1476
- Support internvl2 grounding by @tastelikefeet in #1473
- update doc by @tastelikefeet in #1477
- Support alignment algorithm for vision MLLM by @hjh0119 in #1474
- fix doc by @hjh0119 in #1481
- Fix visual cpo by @hjh0119 in #1482
- support llama3.1-quant by @Jintao-Huang in #1478
- fix part tuner by @tastelikefeet in #1483
- fix import by @tastelikefeet in #1488
- Fix GLM4V by @tastelikefeet in #1490
- support mistral large by @Jintao-Huang in #1485
- fix resume_only_model & zero3 & full by @Jintao-Huang in #1497
- Fix resume_from_checkpoint & full by @Jintao-Huang in #1498
- fix part tuner by @tastelikefeet in #1495
- fix cogvlm2-video by @Jintao-Huang in #1500
- [TorchAcc] add script for qwen2 in torchacc by @Zhikaiiii in #1492
- Fix CI by @tastelikefeet in #1501
- fix vlm template by @Jintao-Huang in #1503
- fix internvl-4b by @Jintao-Huang in #1505
- support zero3 & freeze by @Jintao-Huang in #1508
- fix part mix with lora by @tastelikefeet in #1509
- fix docs by @Jintao-Huang in #1511
- Update README.md by @ArtificialZeng in #1516
- fix kto custom data by @hjh0119 in #1515
- Fix KTO doc by @hjh0119 in #1517
- Rescale image by @tastelikefeet in #1512
- fix pretrain dataset by @Jintao-Huang in #1518
- fix deepseek-vl template by @Jintao-Huang in #1521
- Support exporting of llama3.1, and awq-batch-size by @tastelikefeet in #1522
- support lmdeploy & vlm by @Jintao-Huang in #1364
- fix tf 4.43 llava by @tastelikefeet in #1525
- fix llamapro by @tastelikefeet in #1527
- fix template & docs by @Jintao-Huang in #1529
- fix lmdeploy & vlm by @Jintao-Huang in #1530
- update doc by @tastelikefeet in #1531
- fix lmdeploy & minicpm-v-2.5 by @Jintao-Huang in #1534
- fix internvl-phi3 batch infer by @hjh0119 in #1539
- Support SA1B series datasets by @tastelikefeet in #1542
- fix bug in _prepare_inputs by @guihonghao in #1543
- Support lmdeploy infer deploy by @Jintao-Huang in #1541
- add lmdeploy link by @Jintao-Huang in #1545
- support lmdeploy & app-ui by @Jintao-Huang in #1546
- fix lmdeploy bug by @Jintao-Huang in #1550
- support more models by @Jintao-Huang in #1552
- fix multi node by @tastelikefeet in #1554
- support lmdeploy awq by @Jintao-Huang in #1555
- support quant_policy by @Jintao-Huang in #1556
- fix xcomposer lora by @Jintao-Huang in #1559
- Update docs by @Jintao-Huang in #1558
- fix minicpm-v by @Jintao-Huang in #1562
- add result_dir paramerter to InferArgument & fix a minor bug by @starxhong in #1561
- fix some bugs in dpo by @hjh0119 in #1565
- Fix bugs 0801 by @Jintao-Huang in #1566
- fix dataset copy by @Jintao-Huang in #1569
- fix qwen-vl-merged lmdeploy by @Jintao-Huang in #1572
New Contributors
- @ArtificialZeng made their first contribution in #1516
- @guihonghao made their first contribution in #1543
Full Changelog: v2.2.3...v2.2.5
v2.2.3
New Features:
- support bnb and ollama export
- suport Q-Galore
New Models:
- numina-math-7b
Bug Fix:
- fix vllm>=0.5.1, TP
- fix internvl2 template
- fix glm4v merge-lora
What's Changed
- fix internvl doc by @hjh0119 in #1394
- Fix link by @Jintao-Huang in #1397
- fix vllm==0.5.1 by @Jintao-Huang in #1404
- [TorchAcc] update accelerate API and add llama3-70B by @baoleai in #1400
- Support Ollama and BNB for export by @tastelikefeet in #1407
- Fix glm4v merge lora by @Jintao-Huang in #1410
- [TorchAcc] fix model download when using TorchAcc distributed training by @baoleai in #1408
- Support padding left by @tastelikefeet in #1414
- Fix ollama export by @tastelikefeet in #1416
- fix web-ui params by @tastelikefeet in #1417
- fix hub_token by @Jintao-Huang in #1420
- Update ms hub token by @Jintao-Huang in #1424
- Add numina math model by @tastelikefeet in #1421
- fix internvl template by @Jintao-Huang in #1433
- Internvl series models update by @hjh0119 in #1426
- fix internvl2 template by @Jintao-Huang in #1436
- Fix bug and make lazydataset more stable by @tastelikefeet in #1438
- Fix llava-hf by @tastelikefeet in #1439
- [WIP]Support Q-Galore by @tastelikefeet in #1440
-
- support deepspeed on ui 2. add tools to client_utils by @tastelikefeet in #1446
- fix read csv (float) by @Jintao-Huang in #1447
- fix dataset by @tastelikefeet in #1448
- update internvl doc by @hjh0119 in #1449
Full Changelog: v2.2.2...v2.2.3
v2.2.2
English Version
Features
- Support lmdeploy for inference
- Support training for Internvl2 Video
- Support argument for LoRA target module in regex format
- Support RTD website
- Support argument of device_max_memory to config device_map memory usage.
New Models
- Support CogVLM2-Video
中文
新功能
- 支持lmdeploy框架的推理能力
- 支持InternVL2模型的视频训练能力
- 支持lora_target_regex参数,用来以正则表达式配置lora模块
- 支持RTD文档网站
- 支持device_max_memory参数来配置device_map的显存使用
新模型
- CogVLM2-Video
What's Changed
- fix bug by @tastelikefeet in #1320
- fix template by @Jintao-Huang in #1321
- support cogvlm2-video by @Jintao-Huang in #1318
- Fix bugs by @tastelikefeet in #1325
- fix web-ui by @tastelikefeet in #1327
- compatible with trl 0.9.6 by @hjh0119 in #1326
- compat with vllm==0.5.1 by @Jintao-Huang in #1329
- Update qrcode by @tastelikefeet in #1332
- fix florence model by @hjh0119 in #1334
- Relaxing requirements for trl by @hjh0119 in #1342
- fix xcomposer2.5 device_map by @Jintao-Huang in #1343
- support generation_info by @Jintao-Huang in #1344
- fix requirements by @Jintao-Huang in #1347
- readthedocs by @tastelikefeet in #1345
- fix sequence parallel get labels by @tastelikefeet in #1352
- fix filelock by @tastelikefeet in #1354
- Add pt command by @tastelikefeet in #1356
- fix generation_info efficiency by @Jintao-Huang in #1359
- fix sh ddp_backend by @Jintao-Huang in #1360
- support LLM & lmdeploy by @Jintao-Huang in #1272
- fix a file path by @tastelikefeet in #1363
- Internvl2 support video by @tastelikefeet in #1366
- fix openai api by @Jintao-Huang in #1367
- fix internvl2-40b by @hjh0119 in #1369
- fix vlm deploy lora & agent by @Jintao-Huang in #1371
- Support lora regex by @tastelikefeet in #1375
- Fix docs by @Jintao-Huang in #1380
- Fix FSDP; Add training percentage to jsonl logging; Add a web-ui component by @tastelikefeet in #1381
- Support max memory args by @tastelikefeet in #1382
- fix max_memory by @tastelikefeet in #1383
- Fix gpu assert calculation by @tastelikefeet in #1384
- fix dataset_sample & deploy stop_words by @Jintao-Huang in #1385
Full Changelog: v2.2.1...v2.2.2
v2.2.1
English Version
New Features
- Multimodal: Supported a large number of multimodal datasets and restructured the multimodal architecture. Some models now support grounding tasks.
- Web-ui: Added support for RLHF, evaluation, and quantization.
- Evaluation Functionality: Refactored the evaluation functionality, now using OpenCompass internally, supporting over 50 evaluation datasets.
- Deployment Functionality: VLLM infer_backend now supports multimodal models.
- Agent Training: Refactored the construction, training, and deployment of agent datasets, making the agent pipeline more complete.
- Human Alignment: Added alignment algorithms such as KTO and CPO, and refactored the human alignment code.
New Models
- openbuddy-llama3-70b
- Deepseek-coder-v2
- llava1.5, llava1.6, llava-next-video
- gemma2
- Florence
- phi3-4k
- internlm2.5, xcomposer2.5
- internvl2
- codegeex4
- mistral-7b-instruct-v0.3
New Datasets
- Over 30 foundational multimodal datasets, including GQA, RefCOCO, and Llava-pretrain.
- Swift-mix general mixed dataset.
- Video-chatgpt video dataset.
中文版本
新功能
- 多模态:支持了非常多的多模态数据集,并重构了多模态架构,部分模型开始支持grounding任务
- Web-ui:支持了RLHF、评测和量化
- 评测功能:进行了重构,内部使用了OpenCompass,支持50+评测集
- 部署功能:VLLM infer_backend支持多模态模型
- Agent训练:重构了Agent数据集构造、训练、部署,Agent链路更加完整
- 人类对齐:增加了KTO、CPO等对齐算法,并重构了人类对齐的代码
新模型
- openbuddy-llama3-70b
- Deepseek-coder-v2
- llava1.5, llava1.6, llava-next-video
- gemma2
- Florence
- phi3-4k
- internlm2.5, xcomposer2.5
- internvl2
- codegeex4
- mistral-7b-instruct-v0.3
新数据集
- GQA、RefCOCO、Llava-pretrain等30+多模态基础数据集
- swift-mix通用混合数据集
- video-chatgpt视频数据集
What's Changed
- fix py38 by @Jintao-Huang in #1152
- refactor rlhf by @hjh0119 in #1090
- Fix/web UI 0617 by @tastelikefeet in #1158
- Add more datasets by @tastelikefeet in #1065
- support openbuddy-llama3-70b by @Jintao-Huang in #1162
- Add Deepseek coder v2 model by @tastelikefeet in #1164
- Update web-ui labels by @tastelikefeet in #1168
- fix hf download by @tastelikefeet in #1169
- fix kto bugs by @hjh0119 in #1173
- Fix preprocessor by @tastelikefeet in #1174
- refactor ui by @tastelikefeet in #1178
- fix DPO by @hjh0119 in #1181
- fix bugs by @Jintao-Huang in #1171
- 在argument.py添加hccl参数,用于NPU环境下训练使用 by @jiaozhentian in #1165
- Custom agent dataset training support by @hjh0119 in #1184
- RLHF UI by @tastelikefeet in #1182
- fix kto by @hjh0119 in #1187
- update agent doc by @hjh0119 in #1189
- Fix device map in limited mem by @tastelikefeet in #1190
- Fix sk will be uploaded by @tastelikefeet in #1192
- Fix dataset concatenation by @tastelikefeet in #1193
- Fix glm4v dataloader by @CjangCjengh in #1183
- fix glm4v images by @Jintao-Huang in #1194
- [TorchAcc] Cache the compiled results and remove some xla flags by @anw90 in #1160
- Fix vqa dataset loading by @tastelikefeet in #1195
- [TorchAcc] Add acc_steps arguments to speed up torchacc training by @anw90 in #1196
- fix minicpm-v data_collator by @Jintao-Huang in #1197
- Add a tuner to train a part of the model by @tastelikefeet in #1200
- Fix ConversationPreprocessor loading by @tastelikefeet in #1201
- Fix map cache on multi node by @tastelikefeet in #1203
- refactor multi-modal by @Jintao-Huang in #1212
- support llava 1.5 by @Jintao-Huang in #1217
- add in_browswer by @tastelikefeet in #1220
- [TorchAcc] Add USE_TORCH_XLA=0 flag for native swift scripts by @anw90 in #1221
- Fix glm4v batch_size>1 by @Jintao-Huang in #1223
- Add debug log support by @tastelikefeet in #1226
- Fix qlora deploy by @Jintao-Huang in #1224
- Add new dataset by @tastelikefeet in #1227
- fix refcoco by @tastelikefeet in #1232
- Fix bugs by @tastelikefeet in #1241
- fix bugs by @Jintao-Huang in #1242
- refactor inference by @Jintao-Huang in #1245
- Add more datasets by @tastelikefeet in #1246
- Support gemma2 by @Jintao-Huang in #1247
- Support gemma2 by @Jintao-Huang in #1248
- add push to hub tracker by @tastelikefeet in #1249
- support Florence series model by @hjh0119 in #1244
- update eval by @Jintao-Huang in #1251
- reset snapshot_download to modelscope by @tastelikefeet in #1254
- Feat/open compass by @tastelikefeet in #1213
- fix output empty will be filtered by @tastelikefeet in #1256
- add phi3 models by @tastelikefeet in #1259
- update dataset preprocess by @Jintao-Huang in #1257
- fix rlhf by @tastelikefeet in #1261
- fix rlhf by @Jintao-Huang in #1260
- Support llava1.6 mistral by @Jintao-Huang in #1268
- Add new dataset by @tastelikefeet in #1270
- support vllm & multimodal by @Jintao-Huang in #1229
- Fix bugs by @Jintao-Huang in #1275
- fix llamapro by @tastelikefeet in #1276
- fix xcomposer deploy by @Jintao-Huang in #1280
- Fix/0703 2 by @tastelikefeet in #1284
- Fix llamapro merge and quantized training by @tastelikefeet in #1289
- update lr_scheduler by @Jintao-Huang in #1288
- support internlm2.5 by @Jintao-Huang in #1291
- fix vllm by @Jintao-Huang in #1292
- fix florence & deploy by @Jintao-Huang in #1293
- fix mini-internvl-phi3 template by @Jintao-Huang in #1298
- fix florence by @hjh0119 in #1301
- Refactor UI and fix some bugs by @tastelikefeet in #1300
- fix dependency by @tastelikefeet in #1306
- support codegeex4 by @Jintao-Huang in #1305
- support internvl2 by @hjh0119 in #1304
- support llava video by @Jintao-Huang in #1307
- fix docs by @Jintao-Huang in #1309
- support lr_scheduler_kwargs by @Jintao-Huang in #1310
- Fix internvl2 template by @hjh0119 in #1308
- Fix bugs by @tastelikefeet in #1311
- support warmup_stable_decay by @Jintao-Huang in #1312
- Support xcomposer2.5 by @Jintao-Huang in #1287
- Fix bugs by @tastelikefeet in #1319
New Contributors
- @jiaozhentian made their first contribution in #1165
- @CjangCjengh made their first contribution in #1183
- @anw90 made their first contribution in #1160
Full Changelog: v2.1.1...v2.2.1
v2.1.0
中文
新功能
- 支持了SimPO对齐算法,请查看文档
- 支持多模态部署能力
- web-ui支持量化和评测功能,并支持多模态界面推理和部署
新模型
- ChatGLM4和ChatGLMv
- Qwen2系列
- llava1.5/1.6系列模型
- mini-internvl系列模型
- paligemma系列模型
- Yuan2模型
Bug修复
请查看下方的详细提交记录
English
New Features
- Add SimPO alignment algorithm. Please refer to the documentation.
- Support for multimodal deployment capabilities.
- Web UI now supports quantization and evaluation command, as well as multimodal inference and deployment.
New Models
- ChatGLM4 and ChatGLMv
- Qwen2 series
- llava1.5/1.6 series models
- mini-internvl series models
- paligemma series models
- Yuan2 model
Bug fixing
Please check the update logs for details
What's Changed
- [TorchAcc][Experimental] Integrate more model in torchacc by @Zhikaiiii in #683
- update custom_val_dataset by @Jintao-Huang in #991
- fix val_dataset by @Jintao-Huang in #992
- fix args by @Jintao-Huang in #996
- fix pt deploy lora by @Jintao-Huang in #999
- Feat/phi3 paligemma by @tastelikefeet in #998
- Add copy to template encode by @tastelikefeet in #1003
- 更新了Command-line-parameters.md里面的一个链接 by @KPCOFGS in #1001
- fix cogvlm2 history by @Jintao-Huang in #1005
- fix minicpm-v by @Jintao-Huang in #1010
- Fix gradio conccurency limit by @tastelikefeet in #1016
- Update code by @Jintao-Huang in #962
- support paligemma by @tastelikefeet in #1004
- update docs table by @Jintao-Huang in #1021
- Fix docs table by @Jintao-Huang in #1024
- fix docs and a bug by @tastelikefeet in #1023
- fix arguments by @Jintao-Huang in #1026
- fix custom by @Jintao-Huang in #1028
- fix vllm==0.4.* slower than vllm==0.3.* by @Jintao-Huang in #1035
- fix app_ui host & port by @Jintao-Huang in #1036
- fix bugs by @Jintao-Huang in #1038
- support mini-internvl by @hjh0119 in #1032
- support multimodal deploy by @Jintao-Huang in #1029
- Support SimPO Algorithm by @hjh0119 in #1037
- fix phi3-vision bug by @Jintao-Huang in #1039
- update arguments by @Jintao-Huang in #1043
- update arguments by @Jintao-Huang in #1044
- fix deepseek-vl by @Jintao-Huang in #1046
- support sharegpt dataset format by @Jintao-Huang in #1052
- fix mini-internvl-4b by @hjh0119 in #1053
- Fix mini-internvl-4b by @hjh0119 in #1054
- fix vllm==0.4.3 by @Jintao-Huang in #1055
- fix deepseek-vl template by @Jintao-Huang in #1057
- Fix citest test_run.py by @Jintao-Huang in #1059
- ✏️ fix typo by @bertjiazheng in #1064
- fix argument by @hjh0119 in #1066
- Support glm4 by @Jintao-Huang in #1069
- update wechat by @Jintao-Huang in #1071
New Contributors
- @Zhikaiiii made their first contribution in #683
- @KPCOFGS made their first contribution in #1001
- @bertjiazheng made their first contribution in #1064
Full Changelog: v2.0.5...v2.1.0