Releases: EricFillion/happy-transformer
Releases · EricFillion/happy-transformer
Version 3.0.0
New Features:
- Deepspeed is now supported for fine-tuning.
- Apple's MPS backend is now automatically used for both training and fine-tuning if detected.
- Evaluating data is now used during fine-tuning to track the fine-tuning progress.
- WandB can now be used to log the results from fine-tuning.
- CSV files are supported for training/evaluating text generation and word prediction models. This makes it easy to isolate cases.
- Push models to Hugging Face's Hub with one command.
- Enable saving models periodically during training.
Breaking changes:
- Preprocesses data is now saved in the Hugging Face's Dataset format rather than in JSON format.
- Dictionary argument inputs for training and evaluating are no longer supported
- Removed adam_beta1, adam_beta2, adam_epsilon and max_grad_norm learning parameters.
- Replaced save_preprocessed_data, save_preprocessed_data_path with a single parameter called save_path. Likewise for load_preprocessed_data and load_preprocessed_data_path being replaced by load_path.
- Removed support for dictionary settings for the args parameter for training and evaluating.
- Removed the preprocessing_processes parameter for training and evaluating.
Updated Save Strategy
Contains fix from #280 to ensure that models are not saved during training.
Half Precession Training and Encoding Format
Fixed use_auth_token not being passed to the HappyTransformer class
- Contains fix from #268
- Added text-to-text articles to docs to provide additional examples.
Added Support For Private Models
Allows users to use their authentication token to access their private models from Hugging Face's Model Hub.
See #266
Added Support to Prevent Bad Words/Phrases
2.3.0: Text-to-Text Fine-Tuning
2.2.5: Increase tqdm version
Text-to-text!
- Added text-to-text functionality
- Fixed a bug regarding using word prediction with ALBERT. Sometimes the model would predict a blank string which would cause an exception during post-processing. Also added the preventative code to RoBERTa's post-processing method.
Added Top-p To Text Generation Settings
Merge pull request #235 from EricFillion/ef/fix-gen-setting-table Fixed Text Generation Setting Table