Releases: modelscope/ms-swift
Releases · modelscope/ms-swift
v3.0.1
中文版
新特性
- 支持
SequenceClassification
模型的训练、推理和部署。可以查看以下例子:qwen2.5,bert。 - LlamaPro支持多模态模型. 例如:qwen2vl、internvl2.5、llama3-vision等。
新模型
- Qwen/QVQ-72B-Preview
- iic/DocOwl2
- OpenGVLab/InternVL2-Pretrain-Models, OpenGVLab/InternVL2_5-4B-AWQ系列, OpenGVLab/InternVL2_5-1B-MPO系列
- deepseek-ai/DeepSeek-V3系列
- answerdotai/ModernBERT-base系列
- AI-ModelScope/paligemma2-3b-pt-224系列, AI-ModelScope/paligemma2-3b-ft-docci-448系列
- AI-ModelScope/Skywork-o1-Open-Llama-3.1-8B
English Version
New Features:
- Support for training, inference, and deployment of
SequenceClassification
models. You can check the following examples: qwen2.5, bert. - LlamaPro supports multimodal models, such as qwen2vl, internvl2.5, and llama3-vision.
New Models:
- Qwen/QVQ-72B-Preview
- iic/DocOwl2
- OpenGVLab/InternVL2-Pretrain-Models, OpenGVLab/InternVL2_5-4B-AWQ series, OpenGVLab/InternVL2_5-1B-MPO series
- deepseek-ai/DeepSeek-V3 series
- answerdotai/ModernBERT-base series
- AI-ModelScope/paligemma2-3b-pt-224 series, AI-ModelScope/paligemma2-3b-ft-docci-448 series
- AI-ModelScope/Skywork-o1-Open-Llama-3.1-8B
What's Changed
- Fix mplug owl2, molmo by @Jintao-Huang in #2724
- fix batch_infer pad_token & florence by @Jintao-Huang in #2725
- Support qwen agent format by @tastelikefeet in #2722
- Support more internvl2.5 awq/mpo & internvl2 pretrain model by @Jintao-Huang in #2726
- support iic/DocOwl2 by @Jintao-Huang in #2728
- update examples by @Jintao-Huang in #2730
- remove files by @Jintao-Huang in #2732
- support paligemma2 by @Jintao-Huang in #2735
- fix windows by @Jintao-Huang in #2733
- support multi-modal llamapro by @tastelikefeet in #2738
- support AI-ModelScope/Skywork-o1-Open-Llama-3.1-8B by @Jintao-Huang in #2739
- Fix windows encoding gbk by @Jintao-Huang in #2741
- fix docs multimodal by @Jintao-Huang in #2742
- support SequenceClassification & update QVQ-72B-Preview by @Jintao-Huang in #2747
- fix web-ui by @Jintao-Huang in #2758
- fix bugs by @Jintao-Huang in #2761
- fix shell by @Jintao-Huang in #2764
- fix app-ui by @tastelikefeet in #2765
- support modern_bert & support bert deploy by @Jintao-Huang in #2767
- fix alpaca by @Jintao-Huang in #2771
- support txt by @Jintao-Huang in #2772
- fix telechat2 template by @Jintao-Huang in #2775
- Fix deepspeed by @Jintao-Huang in #2778
- fix qwen2vl by @Jintao-Huang in #2779
- Fix app ui by @tastelikefeet in #2780
- support deepseek-v3 by @Jintao-Huang in #2781
- Fix app-ui by @tastelikefeet in #2784
Full Changelog: v3.0.0...v3.0.1
v3.0.0
中文版
架构修改与新特性:
具体可以查看这里: https://swift.readthedocs.io/zh-cn/latest/Instruction/ReleaseNote3.0.html
新模型:
- OpenGVLab/InternVL2_5-1B等系列模型
- LLM-Research/Llama-3.3-70B-Instruct
- BAAI/Emu3-Gen
- deepseek-ai/DeepSeek-V2.5-1210, deepseek-ai/deepseek-vl2等系列模型
- Shanghai_AI_Laboratory/internlm-xcomposer2d5-ol-7b
- InfiniAI/Megrez-3b-Instruct, InfiniAI/Megrez-3B-Omni
- TeleAI/TeleChat2-3B等系列模型
English Version
Architecture Modifications and New Features:
For more details, please visit: https://swift.readthedocs.io/en/latest/Instruction/ReleaseNote3.0.html
New Models:
- OpenGVLab/InternVL2_5-1B series models
- LLM-Research/Llama-3.3-70B-Instruct
- BAAI/Emu3-Gen
- deepseek-ai/DeepSeek-V2.5-1210, deepseek-ai/deepseek-vl2 series models
- Shanghai_AI_Laboratory/internlm-xcomposer2d5-ol-7b
- InfiniAI/Megrez-3b-Instruct, InfiniAI/Megrez-3B-Omni
- TeleAI/TeleChat2-3B series models
What's Changed
- Refactor All Codes and bump version to 3.0 by @tastelikefeet in #2030
- fix doc by @tastelikefeet in #2545
- fix manifest by @tastelikefeet in #2546
- add doc 2.x by @tastelikefeet in #2548
- fix ui by @tastelikefeet in #2549
- fix infer by @tastelikefeet in #2550
- Refactor mllm by @Jintao-Huang in #2543
- fix ui by @tastelikefeet in #2552
- Fix ui by @tastelikefeet in #2556
- Update ddp infer doc by @Jintao-Huang in #2557
- fix docs by @Jintao-Huang in #2558
- Fix docs by @Jintao-Huang in #2561
- fix log by @tastelikefeet in #2564
- Fix the command line parameter doc by @Jintao-Huang in #2565
- fix context by @Jintao-Huang in #2568
- Documents Updates by @yrk111222 in #2574
- Revert "Documents Updates" by @Jintao-Huang in #2576
- fix hub param by @tastelikefeet in #2572
- Fix bugs by @Jintao-Huang in #2573
- Support internvl2.5 by @Jintao-Huang in #2575
- update english docs by @Jintao-Huang in #2577
- fix en docs by @Jintao-Huang in #2580
- fix docs & add custom example by @Jintao-Huang in #2581
- fix custom example by @Jintao-Huang in #2582
- support llama3.3 by @Jintao-Huang in #2584
- update acc_strategy & fix citest by @Jintao-Huang in #2583
- Support peft0.14 by @tastelikefeet in #2587
- update infer/deploy examples by @Jintao-Huang in #2588
- add image images mapping by @Jintao-Huang in #2594
- update llm sft notebook by @Jintao-Huang in #2599
- fix notebook by @Jintao-Huang in #2600
- Fix streaming by @Jintao-Huang in #2601
- Emu3 gen train by @mi804 in #2602
- compat mllm notebook by @Jintao-Huang in #2604
- Temporarily remove torchacc. by @Jintao-Huang in #2606
- update docs by @Jintao-Huang in #2607
- train and infer scripts for emu3_gen by @mi804 in #2610
- Uodate Document by @yrk111222 in #2615
- update memory usage of emu3-gen by @mi804 in #2611
- move prepare_model by @Jintao-Huang in #2614
- Update mllm notebook by @Jintao-Huang in #2617
- Support all-embedding / all-norm by @Jintao-Huang in #2619
- fix lmdeploy==0.5.* by @Jintao-Huang in #2621
- Support deepseek-ai/DeepSeek-V2.5-1210 by @Jintao-Huang in #2624
- fix use_reentrant gradient_checkpointing by @Jintao-Huang in #2625
- support reward model by @Jintao-Huang in #2628
- fix add_default_tag by @Jintao-Huang in #2631
- fix dataset by @Jintao-Huang in #2636
- fix bugs & update openbuddy models & update docs by @Jintao-Huang in #2638
- fix app-ui by @tastelikefeet in #2641
- Fix post encode by @Jintao-Huang in #2643
- fix bugs by @Jintao-Huang in #2645
- update truncation_strategy by @Jintao-Huang in #2647
- fix swift/Infinity-Instruct by @Jintao-Huang in #2651
- Support LoRA-GA by @lxline in #2650
- support deepseek_vl2 by @Jintao-Huang in #2654
- fix swift/SlimOrca by @Jintao-Huang in #2656
- fix swift/SlimOrca by @Jintao-Huang in #2657
- support Shanghai_AI_Laboratory/internlm-xcomposer2d5-ol-7b:audio by @Jintao-Huang in #2658
- support Shanghai_AI_Laboratory/internlm-xcomposer2d5-ol-7b:base by @Jintao-Huang in #2660
- fix hub by @tastelikefeet in #2661
- fix liger by @tastelikefeet in #2666
- support megrez by @Jintao-Huang in #2667
- fix unsloth resume training by @tastelikefeet in #2668
- fix dataset by @Jintao-Huang in #2670
- Fix bugs by @tastelikefeet in #2671
- fix deepseek_vl2 by @Jintao-Huang in #2675
- support adapters by @Jintao-Huang in #2633
- Support megrez omni by @Jintao-Huang in #2674
- fix docs by @Jintao-Huang in #2679
- fix megrez_omni by @Jintao-Huang in #2680
- fix infer by @Jintao-Huang in #2681
- Fix bugs by @Jintao-Huang in #2687
- Update readme by @Jintao-Huang in #2579
- update wechat by @Jintao-Huang in #2694
- fix readme by @Jintao-Huang in #2696
- Fix web-ui by @tastelikefeet in #2693
- Fix readme by @Jintao-Huang in #2697
- Update banner by @Jintao-Huang in #2699
- fix use_reentrant by @Jintao-Huang in #2700
- update examples by @Jintao-Huang in #2703
- fix eval strategy by @Jintao-Huang in #2707
- Update FAQ by @slin000111 in #2706
- qwen to Qwen by @Jintao-Huang in #2708
- fix timeout & web-ui by @Jintao-Huang in #2709
- Fix multi lora by @tastelikefeet in #2711
- support Qwen/QVQ-72B-Preview by @Jintao-Huang in #2712
- update examples by @Jintao-Huang in #2714
- fix deploy request_config by @Jintao-Huang in #2718
- fix examples by @Jintao-Huang in #2719
- fix gptq group_size by @Jintao-Huang in #2720
- Better error messages by @Jintao-Huang in #2721
New Contributors
- @yrk111222 made their first contribution in #2574
- @lxline made their first contribution in #2650
Full Changelog: v2.6.1...v3.0.0
v2.6.1
New Models:
New Datasets:
What's Changed
- support part tuner replace_key False by @tastelikefeet in #2438
- bump ms version by @tastelikefeet in #2449
- remove useless code by @tastelikefeet in #2453
- fix qwen2-vl position_ids by @Jintao-Huang in #2461
- fix peft is_multimodal by @Jintao-Huang in #2462
- fix qwen2vl pt infer by @Jintao-Huang in #2463
- [TorchAcc] Update padding strategy when using persistent cache by @eedalong in #2464
- fix kto by @Jintao-Huang in #2478
- Update Common QA by @slin000111 in #2475
- fix awq quant device_map by @Jintao-Huang in #2488
- Fix preprocess num proc by @Jintao-Huang in #2492
- Support marco o1 by @Jintao-Huang in #2496
- fix eval_dataset no by @Jintao-Huang in #2497
- support batch flattening collator by @eedalong in #2499
- fix latex-ocr by @Jintao-Huang in #2510
- support mPLUG-Owl3 241101 by @LukeForeverYoung in #2515
- support qwq by @Jintao-Huang in #2520
- support glm-edge & glm-edge-v by @Jintao-Huang in #2526
New Contributors
Full Changelog: v2.6.0...v2.6.1
v2.6.0
English Version
Models
- Support Qwen2.5 coder models
Feature
- Correct and support the new loss and gradient accumulation algorithm from transformers.trainer
中文版本
模型
- 支持千问coder系列模型
功能
- 支持新的transformers loss和GA计算算法,并修正了其中的bug
What's Changed
- fix gptq ovis quant (inputs_embeds) by @Jintao-Huang in #2378
- [TorchAcc] fix qwen2 for transformers>=4.45 by @baoleai in #2379
- fix trl transformers==4.46 compat by @Jintao-Huang in #2385
- fix deploy stream media_type by @Jintao-Huang in #2393
- fix_mplug_owl3_cut_shape by @Jintao-Huang in #2394
- fix swift deploy (lmdeploy stream) by @Jintao-Huang in #2397
- fix lmdeploy warning & gptq-int4 support by @Jintao-Huang in #2401
- support qwen2.5-coder by @Jintao-Huang in #2400
- fix qwen_vl npu by @Jintao-Huang in #2408
- fix model path by @yingdachen in #2410
- fix qwen_vl dpo by @Jintao-Huang in #2411
- Compat transformers 4.46.2 loss by @Jintao-Huang in #2413
- Fix molmo infer by @Jintao-Huang in #2419
- Fix dataset map by @Jintao-Huang in #2421
- Support qwen2 5 coder series model by @Jintao-Huang in #2422
- fix PPO by @hjh0119 in #2377
- fix docs by @Jintao-Huang in #2425
- compat transformers==4.47 by @Jintao-Huang in #2426
- fix internvl2 lmdeploy>=0.6 deploy by @Jintao-Huang in #2429
Full Changelog: v2.5.2...v2.6.0
v2.5.2
New Models:
- emu3-chat
- aya-expanse
- ministral-8b-inst-2410
New Datasets:
- llava-video-178k
- moviechat-1k-test
What's Changed
- fix system by @Jintao-Huang in #2307
- Support Llava-video-178k dataset, a video dataset by @DaozeZhang in #2309
- update molmo readme by @mi804 in #2313
- fix molmo device_map by @Jintao-Huang in #2311
- support emu3-chat by @mi804 in #2322
- Support Moviechat-1k-test dataset, a video datatset by @DaozeZhang in #2319
- fix custom vllm eval args by @Yunnglin in #2325
- fix enable_cache by @Jintao-Huang in #2332
- feat: Add aya models by @Aunali321 in #2335
- compat transformers 4.46 by @Jintao-Huang in #2341
- fix streaming by @tastelikefeet in #2350
- fix doc by @tastelikefeet in #2355
- fix internvl2 eval by @Jintao-Huang in #2362
- fix transformers>=4.46 loss by @Jintao-Huang in #2365
- Support Ministral-8b-inst-2410 by @DaozeZhang in #2371
- fix custom loss by @Jintao-Huang in #2374
New Contributors
- @Aunali321 made their first contribution in #2335
Full Changelog: v2.5.1...v2.5.2
v2.5.1
English Version
New Features:
- Support for RM for LLM and MLLM, as well as PPO for LLM.
New Models:
- molmo series
- mplug-owl3 1b/2b
- llama3.1-nemotron-70b-instruct
- deepseek-janus
中文版
新特性:
- 支持LLM和MLLM的RM, 以及LLM的PPO.
新模型:
- molmo系列
- mplug-owl3 1b/2b
- llama3.1-nemotron-70b-instruct
- deepseek-janus
What's Changed
- support reward modeling and ppo by @hjh0119 in #2093
- fix rescale_image by @tastelikefeet in #2223
- fix deploy timeout by @Jintao-Huang in #2230
- Fix qwen2 vl batch size by @Jintao-Huang in #2239
- Fix ovis1.6 infer by @Jintao-Huang in #2242
- fix publish by @Jintao-Huang in #2245
- fix qwen2vl video args by @Jintao-Huang in #2251
- Update FAQ by @slin000111 in #2252
- Support molmo series vlm by @mi804 in #2260
- fix sft system by @Jintao-Huang in #2262
- support mplug3 1b/2b by @Jintao-Huang in #2271
- Fix deploy openai by @Jintao-Huang in #2278
- fix vllm ignore suffix by @Jintao-Huang in #2287
- fix lora_target_modules in PPO by @hjh0119 in #2274
- fix quant blocks by @Jintao-Huang in #2292
- Support Llama3.1-nemotron-70b-inst-hf by @DaozeZhang in #2299
- fix ppo citest by @hjh0119 in #2302
- support deepseek-janus by @Jintao-Huang in #2300
- update molmo by @Jintao-Huang in #2305
New Contributors
Full Changelog: v2.5.0...v2.5.1
v2.5.0
English Version
New Features:
- Support for GPTQ & AWQ quantization of multimodal LLMs.
- Support for dynamic addition of gradient checkpointing in the ViT section to reduce memory consumption.
- Support for multimodal model pre-training.
New Models:
- llama3.2, llama3.2-vision series
- got-ocr2
- llama3.1-omni
- ovis1.6-gemma2
- pixtral-12b
- telechat2-115b
- mistral-small-inst-2409
New Datasets:
- egoschema
中文版
新特性:
- 支持多模态LLM的gptq&awq量化.
- 支持动态在vit部分增加gradient_checkpointing, 减少显存消耗.
- 支持多模态模型预训练.
新模型:
- llama3.2, llama3.2-vision系列
- got-ocr2
- llama3.1-omni
- ovis1.6-gemma2
- pixtral-12b
- telechat2-115b
- mistral-small-inst-2409
新数据集:
- egoschema
What's Changed
- fix win32 quote by @tastelikefeet in #2065
- Fix yi template by @Jintao-Huang in #2067
- fix rlhf zero3 by @Jintao-Huang in #2072
- Update qwen2-vl最佳实践.md by @Digital2Slave in #2058
- fix RLHF & max_length by @Jintao-Huang in #2075
- Support Mistral-small-inst-2409 by @DaozeZhang in #2077
- dynamic vit gradient_checkpointing by @Jintao-Huang in #2071
- fix qwen2.5 template by @Jintao-Huang in #2081
- fix multiprocess remove_columns by @Jintao-Huang in #2088
- Support for fine-tuning Pixtral-12B. by @Jintao-Huang in #2090
- fix vllm tokenizer by @Jintao-Huang in #2099
- Fix the issue with media_offset in owl3 when batch_size > 1. by @LukeForeverYoung in #2100
- fix deploy openai compat by @Jintao-Huang in #2101
- fix dataset preprocess by @Jintao-Huang in #2102
- fix cpu infer device_map by @Jintao-Huang in #2103
- fix infer device_map by @Jintao-Huang in #2105
- Support for fine-tuning Llama 3.1 Omni. by @Jintao-Huang in #2106
- support vllm & qwen2-vl video by @Jintao-Huang in #2110
- Fix qwen2-vl zero2/3 by @Jintao-Huang in #2114
- fix qwen2-audio by @Jintao-Huang in #2116
- [TorchAcc] fix: fix find_labels and can_return_loss by @baoleai in #2120
- support got-ocr2 by @Jintao-Huang in #2123
- Support for fine-tuning and deployment of the Llama 3.2 series models. by @Jintao-Huang in #2130
- Support fine-tuning MLLama. by @Jintao-Huang in #2132
- fix not impl bug by @Jintao-Huang in #2134
- Compat vllm & qwen2-vl by @Jintao-Huang in #2136
- fix requirements by @Jintao-Huang in #2137
- fix model_type by @Jintao-Huang in #2138
- fix deploy vllm by @Jintao-Huang in #2141
- fix docs by @Jintao-Huang in #2142
- Fix VLM lora by @tastelikefeet in #2140
- support mllm pt by @Jintao-Huang in #2146
- [TorchAcc] fix: fix save config and additional file for swift and peft by @baoleai in #2149
- update quant_device_map by @Jintao-Huang in #2154
- fix qwen2-audio by @Jintao-Huang in #2157
- fix template by @Jintao-Huang in #2160
- compat trl==0.11 by @Jintao-Huang in #2169
- Support for Egoschema, a new video dataset by @DaozeZhang in #2173
- Update FAQ by @slin000111 in #2165
- fix mplug-owl3 infer by @Jintao-Huang in #2175
- Support quant mllm by @Jintao-Huang in #2177
- update setup.py by @Jintao-Huang in #2205
- fix bugs by @Jintao-Huang in #2207
- support telechat2 by @Jintao-Huang in #2210
- Support ovis 1.6 by @Jintao-Huang in #2211
New Contributors
- @Digital2Slave made their first contribution in #2058
- @LukeForeverYoung made their first contribution in #2100
Full Changelog: v2.4.2...v2.5.0
v2.4.2
English Version
New Features:
- RLHF reconstruction, supporting all integrated multimodal models, compatible with DeepSpeed Zero2/Zero3, and supports lazy_tokenize.
- Using infer_backend vllm, inference deployment of multimodal large models supports multiple images.
New Models:
- Qwen2.5 series, Qwen2-vl-72b series (base/instruct/gptq-int4/gptq-int8/awq)
- Qwen2.5-math, Qwen2.5-coder series (base/instruct)
- Deepseek-v2.5
New Datasets:
- longwriter-6k-filtered
中文版
新特性:
- RLHF重构,支持所有已接入的多模态模型,兼容deepspeed zero2/zero3,支持lazy_tokenize
- 使用infer_backend vllm,推理部署多模态大模型支持多图.
新模型:
- qwen2.5系列、qwen2-vl-72b系列(base/instruct/gptq-int4/gptq-int8/awq)
- qwen2.5-math, qwen2.5-coder系列(base/instruct)
- deepseek-v2.5
新数据集:
- longwriter-6k-filtered
What's Changed
- fix model_mapping by @Jintao-Huang in #1982
- fix patch by @Jintao-Huang in #1997
- fix by @tastelikefeet in #1995
- Support Deepseek 2.5 by @DaozeZhang in #1992
- fix EngineGenerationConfig importError of lmdeploy by @irexyc in #1990
- compat lmdeploy==0.6 by @Jintao-Huang in #2001
- Fix rlhf ref model by @Jintao-Huang in #2003
- Support llava1.6-llama3.1-8b-instruct by @DaozeZhang in #2005
- fix lmdeploy qwen_vl by @Jintao-Huang in #2009
- Add FAQ Document by @slin000111 in #2013
- Florence use _post_encode & template support encoder-decoder by @Jintao-Huang in #2019
- refactor rlhf by @Jintao-Huang in #1975
- update code by @Jintao-Huang in #2028
- fix deploy eval kill by @Jintao-Huang in #2029
- Fix olora and pissa saving files which will cause the second saving failed by @tastelikefeet in #2032
- fix rlhf & zero3 by @Jintao-Huang in #2034
- Add longwriter filtered dataset by @wangxingjun778 in #2037
- fix mplug-owl3 by @Jintao-Huang in #2042
- support multi bbox grounding by @tastelikefeet in #2045
- Fix multi coordinate grounding by @tastelikefeet in #2047
- llama3 tool calling by @tastelikefeet in #2048
- update docs by @Jintao-Huang in #2050
- fix qwen2vl position_ids by @Jintao-Huang in #2051
- support qwen2-vl-base by @Jintao-Huang in #2052
- Support qwen2.5 by @Jintao-Huang in #2054
- support qwen2-vl -72b/qwen2.5-math/qwen2.5-coder by @Jintao-Huang in #2056
- vllm support mutli image by @Jintao-Huang in #2059
- support qwen2.5-coder by @Jintao-Huang in #2061
- fix notebook gradio by @Jintao-Huang in #2062
- update qwen2-vl docs by @Jintao-Huang in #2063
New Contributors
- @irexyc made their first contribution in #1990
- @wangxingjun778 made their first contribution in #2037
Full Changelog: v2.4.1...v2.4.2
v2.4.1
English Version
New Features:
- Inference and deployment support for logprobs.
- RLHF support for lazy_tokenize.
- Multimodal model support for neftune.
- dynamic_eos compatibility with glm4 series and other models.
New Models:
- mplug-owl3, best practices can be found here.
- yi-coder 1.5b, base/chat model of 9b.
- minicpm3-4b.
- reflection-llama3.1-70b.
中文版
新功能:
- 推理和部署支持 logprobs。
- RLHF支持lazy_tokenize。
- 多模态模型支持neftune。
- dynamic_eos兼容glm4系列等模型。
新模型:
- mplug-owl3,最佳实践可以查看这里。
- yi-coder 1.5b、9b 的base/chat模型。
- minicpm3-4b。
- reflection-llama3.1-70b。
What's Changed
- Fix push_to_hub when last-checkpoint by @tastelikefeet in #1897
- support custom quantized dataset by @tastelikefeet in #1893
- fix push_to_ms by @tastelikefeet in #1901
- support logprobs by @Jintao-Huang in #1900
- deepspeed use cosine lr_schduler by @Jintao-Huang in #1907
- update docs by @Jintao-Huang in #1908
- fix web-ui push to hub strategy by @tastelikefeet in #1909
- Refactor docs by @tastelikefeet in #1912
- refactor docs by @tastelikefeet in #1915
- [TorchAcc] perf: use xm.save instead of torch.save by @baoleai in #1916
- update wechat by @tastelikefeet in #1925
- update docs & fix bug by @Jintao-Huang in #1926
- [TorchAcc] fix: fix the judegement of fsdp_num by @baoleai in #1903
- Support deploy & logprobs by @Jintao-Huang in #1833
- fix typing by @Jintao-Huang in #1933
- fix swift deploy by @Jintao-Huang in #1936
- update yi-coder by @Jintao-Huang in #1942
- fix lmdeploy seed by @Jintao-Huang in #1945
- fix do_sample by @Jintao-Huang in #1946
- refactor rlhf by @hjh0119 in #1885
- fix file rename error in megatron when there are multi process by @Zhikaiiii in #1948
- fix qwen2-vl & video by @Jintao-Huang in #1950
- support dynamic_eos by @Jintao-Huang in #1947
- fix rlhf by @hjh0119 in #1949
- Support minicpm 3 by @Jintao-Huang in #1952
- Add lazy_tokenize to RLHF by @tastelikefeet in #1956
- Fix data info print in rlhf by @tastelikefeet in #1964
- Fix the lora hook by @tastelikefeet in #1963
- fix bugs by @Jintao-Huang in #1959
- support mplug_owl3 by @Jintao-Huang in #1957
- update docs by @Jintao-Huang in #1970
- Add reflection model by @tastelikefeet in #1973
- fix typo by @Jintao-Huang in #1980
Full Changelog: v2.4.0...v2.4.1
v2.4.0
English Version
New Features:
- Support for Liger, which accommodates models like LLaMA, Qwen, Mistral, etc., and reduces memory usage by 10% to 60%.
- Support for custom loss function training using a registration mechanism.
- Training now supports pushing models to ModelScope and HuggingFace.
- Support for the
freeze_vit
parameter to control the behavior of full parameter training for multimodal models.
New Models:
- Qwen2-VL series includes GPTQ/AWQ quantized models. For best practices, see here.
- InternVL2 AWQ quantized models.
New Datasets:
- qwen2-pro series
中文版
新特性:
- 支持 Liger训练LLaMA、Qwen、Mistral 等模型,内存使用降低 10% 至 60%。
- 支持使用注册机制进行自定义损失函数的训练。
- 训练支持将模型推送至 ModelScope 和 HuggingFace。
- 支持 freeze_vit 参数,以控制多模态模型全参数训练的行为。
新模型:
- Qwen2-VL 系列包括 GPTQ/AWQ 量化模型,最佳实践可以查看这里。
- InternVL2 AWQ 量化模型。
新数据集:
- qwen2-pro 系列
What's Changed
- compat with vllm==0.5.5 by @Jintao-Huang in #1812
- Support zero2 offload by @Jintao-Huang in #1814
- fix mp+ddp & resume_from_checkpoint by @Jintao-Huang in #1815
- fix preprocess_num_proc by @Jintao-Huang in #1818
- Support liger by @tastelikefeet in #1819
- fix dora deployment by @tastelikefeet in #1821
- Support register loss func by @Jintao-Huang in #1822
- use default-lora by @Jintao-Huang in #1823
- fix minicpm-v 2.6 infer device_map by @Jintao-Huang in #1832
- Fix code by @tastelikefeet in #1824
- fix inject by @tastelikefeet in #1835
- support qwen2-pro dataset by @Jintao-Huang in #1834
- add ddp_timeout parameter by @tastelikefeet in #1836
- fix internlm-xcomposer rlhf by @hjh0119 in #1838
- Support eval_nproc by @tastelikefeet in #1843
- support qwen2-vl by @Jintao-Huang in #1842
- Add internvl2 awq models by @tastelikefeet in #1846
- Fix some datasets for streaming by @tastelikefeet in #1848
- Fix Pissa and OLoRA by @tastelikefeet in #1852
- Support qwen2 vl grounding by @tastelikefeet in #1854
- support qwen2-vl & video finetune by @Jintao-Huang in #1849
- Update new datasets by @tastelikefeet in #1855
- update qwen2-vl docs by @Jintao-Huang in #1856
- update qwen2-vl docs by @Jintao-Huang in #1858
- fix qwen2-vl docs by @Jintao-Huang in #1861
- fix requirements by @Jintao-Huang in #1864
- update docs qwen2-vl by @Jintao-Huang in #1869
- Support faster data map by @tastelikefeet in #1871
- [TorchAcc] fix serveral bugs for torchacc FSDP. by @baoleai in #1872
- Add train record by @tastelikefeet in #1873
- Fix num_proc by @Jintao-Huang in #1874
- Fix neftune doc by @tastelikefeet in #1875
- add duet by @tastelikefeet in #1877
- use model.generation_config by @Jintao-Huang in #1850
- Support freeze vit by @Jintao-Huang in #1880
- support qwen2-vl gptq awq by @Jintao-Huang in #1884
- Refactor push_to_hub by @tastelikefeet in #1883
- Fix push to hub logic by @tastelikefeet in #1888
- add vllm lmdeploy benchmark by @Jintao-Huang in #1889
- Add some warnings and fix RLHF by @tastelikefeet in #1890
Full Changelog: v2.3.2...v2.4.0