Skip to content

Releases: neuralmagic/transformers

NM Transformers 1.7

15 Mar 02:31
4589fc0
Compare
Choose a tag to compare

What's Changed

  • GitHub Actions fixed to enable manual runs.
  • Mistral model architecture was updated to enable quantization of the activations.

NM Transformers 1.6

11 Dec 21:27
da1e43d
Compare
Choose a tag to compare
fix reporting and manual run issues (#97) (#98)

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

NM Transformers v1.5.1

23 Aug 18:42
Compare
Choose a tag to compare

What's Changed

NM Transformers Latest

24 Jan 22:21
46f78c1
Compare
Choose a tag to compare
Pre-release
nightly

Refactor onto transformers v4.23.1

NM Transformers v1.4

03 Jan 16:34
46f78c1
Compare
Choose a tag to compare
NM Transformers v1.4 Pre-release
Pre-release
Upgrade to transformers release V4.23.1 (#62)

* Update trainer and model flows to accommodate sparseml

Disable FP16 on QAT start (#12)

* Override LRScheduler when using LRModifiers

* Disable FP16 on QAT start

* keep wrapped scaler object for training after disabling

Using QATMatMul in DistilBERT model class (#41)

Removed double quantization of output of context layer. (#45)

Fix DataParallel validation forward signatures (#47)

* Fix: DataParallel validation forward signatures

* Update: generalize forward_fn selection

Best model after epoch (#46)

fix sclaer check for non fp16 mode in trainer (#38)

Mobilebert QAT (#55)

* Remove duplicate quantization of vocabulary.

enable a QATWrapper for non-parameterized matmuls in BERT self attention (#9)

* Utils and auxillary changes

update Zoo stub loading for SparseZoo 1.1 refactor (#54)

add flag to signal NM integration is active (#32)

Add recipe_name to file names

* Fix errors introduced in manual cherry-pick upgrade

Co-authored-by: Benjamin Fineran <bfineran@users.noreply.github.com>

NM Transformers v1.3

04 Nov 15:44
46f78c1
Compare
Choose a tag to compare
NM Transformers v1.3 Pre-release
Pre-release
Upgrade to transformers release V4.23.1 (#62)

* Update trainer and model flows to accommodate sparseml

Disable FP16 on QAT start (#12)

* Override LRScheduler when using LRModifiers

* Disable FP16 on QAT start

* keep wrapped scaler object for training after disabling

Using QATMatMul in DistilBERT model class (#41)

Removed double quantization of output of context layer. (#45)

Fix DataParallel validation forward signatures (#47)

* Fix: DataParallel validation forward signatures

* Update: generalize forward_fn selection

Best model after epoch (#46)

fix sclaer check for non fp16 mode in trainer (#38)

Mobilebert QAT (#55)

* Remove duplicate quantization of vocabulary.

enable a QATWrapper for non-parameterized matmuls in BERT self attention (#9)

* Utils and auxillary changes

update Zoo stub loading for SparseZoo 1.1 refactor (#54)

add flag to signal NM integration is active (#32)

Add recipe_name to file names

* Fix errors introduced in manual cherry-pick upgrade

Co-authored-by: Benjamin Fineran <bfineran@users.noreply.github.com>

v1.5

21 Feb 15:35
cc9aacf
Compare
Choose a tag to compare

What's Changed

  • Question answering trainer with SparseML integration by @natuan in #1
  • Use existing config for teacher model by @natuan in #3
  • Export ONNX models with named inputs/outputs, samples by @natuan in #4
  • Load and save SparseML QAT recipes by @natuan in #2
  • Fix input/output shapes for exported samples by @natuan in #5
  • Refactor SparseML trainer for multiple tasks by @natuan in #7
  • enable a QATWrapper for non-parameterized matmuls in BERT self attention by @bfineran in #9
  • Override LRScheduler when using LRModifiers by @bfineran in #11
  • Disable FP16 on QAT start by @bfineran in #12
  • Default behavior for sparseml w/o recipes by @natuan in #13
  • Simplify saved recipes by @natuan in #17
  • Integration of BERT QQP + MNLI(Aka GLUE Tasks) by @spacemanidol in #6
  • Integrate SparseML with Masked LM training by @natuan in #16
  • Use distillation modifier from SparseML for QA task by @natuan in #15
  • Update setup.py by @rahul-tuli in #20
  • updated fix: initialize is_qat_recipe in preprocess_state_dict by @bfineran in #18
  • rm datasets/huggingface-hub versions by @rahul-tuli in #23
  • SparseML integ for NER by @natuan in #21
  • support to download and unwrap framework files from SparseZoo stubs by @bfineran in #19
  • Add preprocessing num workers for GLUE by @natuan in #26
  • changing qa to ner by @spacemanidol in #28
  • Recipe args for NER, GLUE, QA and Masked LM by @natuan in #30
  • Fix incorrect steps calculation when gradient acc. by @eldarkurtic in #31
  • add flag to signal NM integration is active by @bfineran in #32
  • update modifiers import for latest sparseml refactors by @bfineran in #34
  • Update 4.7.0 -> 4.18.0 by @KSGulin in #37
  • fix sclaer check for non fp16 mode in trainer by @bfineran in #38
  • Add log_frequency arg by @KSGulin in #33
  • Using QATMatMul in DistilBERT model class by @natuan in #41
  • Fix distilbert scaling by @anmarques in #43
  • Revert "Fix distilbert scaling (#43)" by @bfineran in #44
  • Removed double quantization of output of context layer. by @anmarques in #45
  • Fix DataParallel validation forward signatures by @KSGulin in #47
  • Best model after epoch by @natuan in #46
  • update Zoo stub loading for SparseZoo 1.1 refactor by @bfineran in #54
  • Mobilebert QAT by @anmarques in #55

New Contributors

Full Changelog: v1.4...v1.5

NM Transformers v1.2

07 Sep 17:29
cc9aacf
Compare
Choose a tag to compare
Mobilebert QAT (#55)

* Remove duplicate quantization of vocabulary.

* Remove duplicate quantization of vocabulary.

NM Transformers v1.1

29 Jun 21:10
15ba9b7
Compare
Choose a tag to compare
Best model after epoch (#46)

NM Transformers v1.0

21 Apr 21:41
15ba9b7
Compare
Choose a tag to compare
NM Transformers v1.0 Pre-release
Pre-release
Best model after epoch (#46)