Skip to content

v1.5

Compare
Choose a tag to compare
@rahul-tuli rahul-tuli released this 21 Feb 15:35
cc9aacf

What's Changed

  • Question answering trainer with SparseML integration by @natuan in #1
  • Use existing config for teacher model by @natuan in #3
  • Export ONNX models with named inputs/outputs, samples by @natuan in #4
  • Load and save SparseML QAT recipes by @natuan in #2
  • Fix input/output shapes for exported samples by @natuan in #5
  • Refactor SparseML trainer for multiple tasks by @natuan in #7
  • enable a QATWrapper for non-parameterized matmuls in BERT self attention by @bfineran in #9
  • Override LRScheduler when using LRModifiers by @bfineran in #11
  • Disable FP16 on QAT start by @bfineran in #12
  • Default behavior for sparseml w/o recipes by @natuan in #13
  • Simplify saved recipes by @natuan in #17
  • Integration of BERT QQP + MNLI(Aka GLUE Tasks) by @spacemanidol in #6
  • Integrate SparseML with Masked LM training by @natuan in #16
  • Use distillation modifier from SparseML for QA task by @natuan in #15
  • Update setup.py by @rahul-tuli in #20
  • updated fix: initialize is_qat_recipe in preprocess_state_dict by @bfineran in #18
  • rm datasets/huggingface-hub versions by @rahul-tuli in #23
  • SparseML integ for NER by @natuan in #21
  • support to download and unwrap framework files from SparseZoo stubs by @bfineran in #19
  • Add preprocessing num workers for GLUE by @natuan in #26
  • changing qa to ner by @spacemanidol in #28
  • Recipe args for NER, GLUE, QA and Masked LM by @natuan in #30
  • Fix incorrect steps calculation when gradient acc. by @eldarkurtic in #31
  • add flag to signal NM integration is active by @bfineran in #32
  • update modifiers import for latest sparseml refactors by @bfineran in #34
  • Update 4.7.0 -> 4.18.0 by @KSGulin in #37
  • fix sclaer check for non fp16 mode in trainer by @bfineran in #38
  • Add log_frequency arg by @KSGulin in #33
  • Using QATMatMul in DistilBERT model class by @natuan in #41
  • Fix distilbert scaling by @anmarques in #43
  • Revert "Fix distilbert scaling (#43)" by @bfineran in #44
  • Removed double quantization of output of context layer. by @anmarques in #45
  • Fix DataParallel validation forward signatures by @KSGulin in #47
  • Best model after epoch by @natuan in #46
  • update Zoo stub loading for SparseZoo 1.1 refactor by @bfineran in #54
  • Mobilebert QAT by @anmarques in #55

New Contributors

Full Changelog: v1.4...v1.5