Releases: foundation-model-stack/fms-hf-tuning
Releases · foundation-model-stack/fms-hf-tuning
v2.0.0-rc.1
What's Changed
- fix: do not resize embedding layer by default by @kmehant in #310
- fix: logger is unbound error by @HarikrishnanBalagopal in #308
- feat: Enable JSON dataset compatibility by @willmj in #297
- doc: How to tune LoRA lm_head by @aluu317 in #305
- docs: Add findings from exploration into model tuning performance degradation by @willmj in #315
- fix: warnings about casing when building the Docker image by @HarikrishnanBalagopal in #318
- fix: need to pass skip_prepare_dataset for pretokenized dataset due to breaking change in HF SFTTrainer by @HarikrishnanBalagopal in #326
- feat: install fms-acceleration to enable qlora by @anhuong in #284
- feat: Migrating the trainer controller to python logger by @seshapad in #309
- fix: remove fire ported from Hari's PR #303 by @HarikrishnanBalagopal in #324
- dep: cap transformers version due to FSDP bug by @anhuong in #335
- deps: Add protobuf to support aLLaM models by @willmj in #336
- fix: add enable_aim build args in all stages needed by @anhuong in #337
- fix: remove lm_head post processing by @Abhishek-TAMU in #333
- doc: Add qLoRA README by @aluu317 in #322
- feat: Add deps to evaluate qLora tuned model by @aluu317 in #312
- feat: Add support for smoothly resuming training from a saved checkpoint by @Abhishek-TAMU in #300
- ci: add a github workflow to label pull requests based on their title by @HarikrishnanBalagopal in #298
- fix: Addition of default pad token in tokenizer when EOS and PAD token are equal by @Abhishek-TAMU in #343
- feat: Add DataClass Arguments to Activate Padding-Free and MultiPack Plugin and FastKernels by @achew010 in #280
- fix: cap transformers at v4.44 by @anhuong in #349
- fix: utilities to post process checkpoint for LoRA by @Ssukriti in #338
- feat: Add post processing logic to accelerate launch by @willmj in #351
- build: install additional fms-acceleration plugins by @anhuong in #350
- fix: unable to find output_dir in multi-GPU during resume_from_checkpoint check by @Abhishek-TAMU in #352
New Contributors
Full Changelog: v1.2.1...v2.0.0-rc.1
v1.2.2
v1.2.1
v1.2.1-rc.1
What's Changed
Full Changelog: v1.2.0...v1.2.1-rc.1
v1.2.0
Dependency Updates
- Update packaging requirement from less than 23.2 to a max of 24 to less than 23.2 to max of 25
API/Interface Changes
- Add optional save_model_dir where final checkpoint is saved. See https://github.com/foundation-model-stack/fms-hf-tuning/blob/main/README.md#saving-checkpoints-while-training
Full List of What's Changed
- Add config_utils tests by @aluu317 in #262
- bug: On save event added to callback by @seshapad in #256
- feat: All metric handling changes by @seshapad in #263
- feat: Configuration to set logging level for trigger log by @seshapad in #241
- Data custom collator by @Ssukriti in #260
- feat: per process state metric by @HarikrishnanBalagopal in #239
- feat: Add a dockerfile argument to enable aimstack by @dushyantbehl in #261
- Set default value of target_modules to be None in LoraConfig by @willmj in #269
- feat: Support pretokenized by @kmehant in #272
- Update packaging requirement from <24,>=23.2 to >=23.2,<25 by @dependabot in #212
- Enabling tests for prompt tuning by @Abhishek-TAMU in #278
- fix: do not add special tokens for custom tokenizer by @kmehant in #279
- fix: bug where the logger was not being used properly by @HarikrishnanBalagopal in #286
- Add functionality to free disk space from Github Actions by @willmj in #287
- Add unit test to verify target_modules defaults correctly by @willmj in #281
- docs: Add documentation on experiment tracking. by @dushyantbehl in #257
- Ensure additional metadata to trackers don't throw error in happy case. by @dushyantbehl in #290
- fix: multiple runid creation bug with distributed training by @dushyantbehl in #268
- feat: logging control operation by @seshapad in #264
- fix run evaluation to get base model path by @anhuong in #273
- Fix: Removal of transformers logger and addition of python native logger by @Abhishek-TAMU in #270
- feat: Added additional events such as on_step_begin, on_optimizer_step, on_substep_end by @seshapad in #293
- Always update setuptools to latest by @jbusche in #288
- Rename all fixtures with correct .jsonl extension by @willmj in #295
- feat: add save_model_dir flag where final checkpoint saved by @anhuong in #291
- feat: Example log controller yaml with training state by @seshapad in #296
New Contributors
Full Changelog: v1.1.0...v1.2.0
v1.2.0-rc.1
What's Changed
- Add config_utils tests by @aluu317 in #262
- bug: On save event added to callback by @seshapad in #256
- feat: All metric handling changes by @seshapad in #263
- feat: Configuration to set logging level for trigger log by @seshapad in #241
- deps: limit peft deps by @anhuong in #274
- Data custom collator by @Ssukriti in #260
- Revert "limit peft deps until investigate (#274)" by @anhuong in #275
- feat: per process state metric by @HarikrishnanBalagopal in #239
- feat: Add a dockerfile argument to enable aimstack by @dushyantbehl in #261
- Set default value of target_modules to be None in LoraConfig by @willmj in #269
- feat: Support pretokenized by @kmehant in #272
- Update packaging requirement from <24,>=23.2 to >=23.2,<25 by @dependabot in #212
- Enabling tests for prompt tuning by @Abhishek-TAMU in #278
- fix: do not add special tokens for custom tokenizer by @kmehant in #279
- fix: bug where the logger was not being used properly by @HarikrishnanBalagopal in #286
- Add functionality to free disk space from Github Actions by @willmj in #287
- Add unit test to verify target_modules defaults correctly by @willmj in #281
- docs: Add documentation on experiment tracking. by @dushyantbehl in #257
- Ensure additional metadata to trackers don't throw error in happy case. by @dushyantbehl in #290
- fix: multiple runid creation bug with distributed training by @dushyantbehl in #268
- feat: logging control operation by @seshapad in #264
- fix run evaluation to get base model path by @anhuong in #273
- Fix: Removal of transformers logger and addition of python native logger by @Abhishek-TAMU in #270
- FIX: Metrics file epoch indexing starting from 0 by @Abhishek-TAMU in #294
- feat: Added additional events such as on_step_begin, on_optimizer_step, on_substep_end by @seshapad in #293
- Always update setuptools to latest by @jbusche in #288
- Rename all fixtures with correct .jsonl extension by @willmj in #295
- feat: add save_model_dir flag where final checkpoint saved by @anhuong in #291
New Contributors
Full Changelog: v1.1.0-rc.1...v1.2.0-rc.1
v1.1.0
What's Changed
- fix: Added correct link in main readme for the trainer-controller readme by @seshapad in #254
- trainer controller doc updates by @alex-jw-brooks in #244
- docs: fix the instructions for running with LORA by @HarikrishnanBalagopal in #265
- refactor code to preprocess datasets by @Ssukriti in #259
- Replace shutil.copytree() to fix permission error by @olson-ibm in #251
- fix: logic for getting tracker config by @HarikrishnanBalagopal in #267
- fix: remove lm_head for granite with llama arch models by @anhuong in #258
Full Changelog: v1.0.0...v1.1.0
v1.1.0-rc.1
What's Changed
- fix: Added correct link in main readme for the trainer-controller readme by @seshapad in #254
- trainer controller doc updates by @alex-jw-brooks in #244
- docs: fix the instructions for running with LORA by @HarikrishnanBalagopal in #265
- refactor code to preprocess datasets by @Ssukriti in #259
- Replace shutil.copytree() to fix permission error by @olson-ibm in #251
- fix: logic for getting tracker config by @HarikrishnanBalagopal in #267
- fix: remove lm_head for granite with llama arch models by @Ssukriti in #258
Full Changelog: v1.0.0-rc.1...v1.1.0-rc.1
v1.0.0
New major features:
- Trainer Controller and ability to stop training earlier using metrics and thresholds. See https://github.com/foundation-model-stack/fms-hf-tuning?tab=readme-ov-file#trainer-controller-framework
- Experiment Tracker while tuning for more insights (choice of frameworks)
Dependency upgrades:
- Transformers to minimum version 4.42.0 needed for tuning of granite with llama models
API /interface changes:
- Introduction of tokenizer field in ModelArguments https://github.com/foundation-model-stack/fms-hf-tuning/blob/main/tuning/config/configs.py#L51 to allow custom tokenizer while tuning
- Deprecation of tokenizer field from PromptTuningConfig and moved to ModelArguments
Full List of what's Changed:
- minor refactor to allow modular functions by @Ssukriti in #224
- feat: support some metrics being 'None' without stopping training by @HarikrishnanBalagopal in #169
- Support custom embedding layer resizing to the desired multiple by @kmehant in #227
- ReIntroduce Package for FMS Accel by @fabianlim in #223
- feat: History based loss metric by @seshapad in #156
- docs: instructions for using the trainer controller framework by @HarikrishnanBalagopal in #214
- feat: support custom tokenizer by @kmehant in #229
- tests: tracker unit tests by @dushyantbehl in #172
- feat: Aim runid export by @dushyantbehl in #90
- Move default operations and metrics to variables by @alex-jw-brooks in #249
- Update README.md by @Ssukriti in #240
- deps: Update transformers to latest and skip broken prompt tuning tests by @Abhishek-TAMU in #246
- deps: Update transformers lower bound version by @Abhishek-TAMU in #250
- V100rc1 release by @olson-ibm in #252
New Contributors
- @Abhishek-TAMU made their first contribution in #246
- @olson-ibm made their first contribution in #252
Full Changelog: v0.4.0...v1.0.0
v1.0.0-rc.1
What's Changed
- minor refactor to allow modular functions by @Ssukriti in #224
- feat: support some metrics being 'None' without stopping training by @HarikrishnanBalagopal in #169
- Support custom embedding layer resizing to the desired multiple by @kmehant in #227
- ReIntroduce Package for FMS Accel by @fabianlim in #223
- feat: History based loss metric by @seshapad in #156
- docs: instructions for using the trainer controller framework by @HarikrishnanBalagopal in #214
- feat: support custom tokenizer by @kmehant in #229
- tests: tracker unit tests by @dushyantbehl in #172
- feat: Aim runid export by @dushyantbehl in #90
- Move default operations and metrics to variables by @alex-jw-brooks in #249
- Update README.md by @Ssukriti in #240
- deps: Update transformers to latest and skip broken prompt tuning tests by @Abhishek-TAMU in #246
- deps: Update transformers lower bound version by @Abhishek-TAMU in #250
New Contributors
- @Abhishek-TAMU made their first contribution in #246
Full Changelog: v0.4.0-rc.3...v1.0.0-rc.1