Skip to content

XTuner Release V0.1.19

Compare
Choose a tag to compare
@pppppM pppppM released this 11 May 09:50
· 51 commits to main since this release
aab528c

What's Changed

  • [Fix] LLaVA-v1.5 official settings by @LZHgrla in #594
  • [Feature] Release LLaVA-Llama-3-8B by @LZHgrla in #595
  • [Improve] Add single-gpu configs for LLaVA-Llama-3-8B by @LZHgrla in #596
  • [Docs] Add wisemodel badge by @LZHgrla in #597
  • [Feature] Support load_json_file with json.load by @HIT-cwh in #610
  • [Feature]Support Mircosoft Phi3 4K&128K Instruct Models by @pppppM in #603
  • [Fix] set dataloader_num_workers=4 for llava training by @LZHgrla in #611
  • [Fix] Do not set attn_implementation to flash_attention_2 or sdpa if users already set it in XTuner configs. by @HIT-cwh in #609
  • [Release] LLaVA-Phi-3-mini by @LZHgrla in #615
  • Update README.md by @eltociear in #608
  • [Feature] Refine sp api by @HIT-cwh in #619
  • [Feature] Add conversion scripts for LLaVA-Llama-3-8B by @LZHgrla in #618
  • [Fix] Convert nan to 0 just for logging by @HIT-cwh in #625
  • [Docs] Delete colab and add speed benchmark by @HIT-cwh in #617
  • [Feature] Support dsz3+qlora by @HIT-cwh in #600
  • [Feature] Add qwen1.5 110b cfgs by @HIT-cwh in #632
  • check transformers version before dispatch by @HIT-cwh in #672
  • [Fix] convert_xtuner_weights_to_hf with frozen ViT by @LZHgrla in #661
  • [Fix] Fix batch-size setting of single-card LLaVA-Llama-3-8B configs by @LZHgrla in #598
  • [Feature] add HFCheckpointHook to auto save hf model after the whole training phase by @HIT-cwh in #621
  • Remove test info in DatasetInfoHook by @hhaAndroid in #622
  • [Improve] Support safe_serialization saving by @LZHgrla in #648
  • bump version to 0.1.19 by @HIT-cwh in #675

New Contributors

Full Changelog: v0.1.18...v0.1.19