XTuner Release V0.1.19
What's Changed
- [Fix] LLaVA-v1.5 official settings by @LZHgrla in #594
- [Feature] Release LLaVA-Llama-3-8B by @LZHgrla in #595
- [Improve] Add single-gpu configs for LLaVA-Llama-3-8B by @LZHgrla in #596
- [Docs] Add wisemodel badge by @LZHgrla in #597
- [Feature] Support load_json_file with json.load by @HIT-cwh in #610
- [Feature]Support Mircosoft Phi3 4K&128K Instruct Models by @pppppM in #603
- [Fix] set
dataloader_num_workers=4
for llava training by @LZHgrla in #611 - [Fix] Do not set attn_implementation to flash_attention_2 or sdpa if users already set it in XTuner configs. by @HIT-cwh in #609
- [Release] LLaVA-Phi-3-mini by @LZHgrla in #615
- Update README.md by @eltociear in #608
- [Feature] Refine sp api by @HIT-cwh in #619
- [Feature] Add conversion scripts for LLaVA-Llama-3-8B by @LZHgrla in #618
- [Fix] Convert nan to 0 just for logging by @HIT-cwh in #625
- [Docs] Delete colab and add speed benchmark by @HIT-cwh in #617
- [Feature] Support dsz3+qlora by @HIT-cwh in #600
- [Feature] Add qwen1.5 110b cfgs by @HIT-cwh in #632
- check transformers version before dispatch by @HIT-cwh in #672
- [Fix]
convert_xtuner_weights_to_hf
with frozen ViT by @LZHgrla in #661 - [Fix] Fix batch-size setting of single-card LLaVA-Llama-3-8B configs by @LZHgrla in #598
- [Feature] add HFCheckpointHook to auto save hf model after the whole training phase by @HIT-cwh in #621
- Remove test info in DatasetInfoHook by @hhaAndroid in #622
- [Improve] Support
safe_serialization
saving by @LZHgrla in #648 - bump version to 0.1.19 by @HIT-cwh in #675
New Contributors
- @eltociear made their first contribution in #608
Full Changelog: v0.1.18...v0.1.19