Skip to content

Releases: TransformerLensOrg/TransformerLens

v2.15.0

20 Feb 00:35
e65fafb
Compare
Choose a tag to compare

Nice little update! This one improves compatibility for LLaMA 3.3 70B, adds a new mistral mode, and introduces a number of utilities for Bert.

What's Changed

New Contributors

Full Changelog: v2.14.1...v2.15.0

v2.14.1

18 Feb 01:37
5e328e9
Compare
Choose a tag to compare

Reenables support for most recent version ofTransformerLens

What's Changed

Full Changelog: v2.14.0...v2.14.1

v2.14.0

13 Feb 01:50
d7f9eb1
Compare
Choose a tag to compare

Much more robust, but still experimental multi-GPU support!

What's Changed

Full Changelog: v2.13.0...v2.14.0

v2.13.0

05 Feb 00:43
53dee84
Compare
Choose a tag to compare

Nice little maintenance one, and a large expansion of generation to allow support for vision models!

What's Changed

New Contributors

Full Changelog: v2.12.0...v2.13.0

v2.12.0

21 Jan 00:07
db0f191
Compare
Choose a tag to compare

What's Changed

Full Changelog: v2.11.0...v2.12.0

v2.11.0

31 Dec 02:04
f103deb
Compare
Choose a tag to compare

LLaMA 3.3 support! This release also includes a handful of usability improvements.

What's Changed

Full Changelog: v2.10.0...v2.11.0

v2.10.0

14 Dec 00:56
30c90f4
Compare
Choose a tag to compare

Huge update! This is likely going to be the last big 2.x update. This update greatly improves model implementation accuracy, and adds some of the newer Qwen models.

What's Changed

New Contributors

Full Changelog: v2.9.1...v2.10.0

v2.9.1

19 Nov 14:34
3267a43
Compare
Choose a tag to compare

Minor dependency change to address a change in an outside dependency

What's Changed

Full Changelog: v2.9.0...v2.9.1

v2.9.0

16 Nov 00:28
dc19c08
Compare
Choose a tag to compare

Lot's of accuracy improvements! A number of models are behaving closer to how they behave in Transformers, and a new internal configuration has been added to allow for more ease of use!

What's Changed

  • fix the bug that attention_mask and past_kv_cache cannot work together by @yzhhr in #772
  • Set prepend_bos to false by default for Bloom model family by @degenfabian in #775
  • Fix that if use_past_kv_cache is set to True models from the Bloom family produce weird outputs. by @degenfabian in #777

New Contributors

Full Changelog: v2.8.1...v2.9.0

v2.8.1

26 Oct 21:12
8f482fc
Compare
Choose a tag to compare

New notebook for comparing models, and bug fix with dealing with newer LLaMA models!

What's Changed

  • Logit comparator tool by @curt-tigges in #765
  • Add support for NTK-by-Part Rotary Embedding & set correct rotary base for Llama-3.1 series by @Hzfinfdu in #764

New Contributors

Full Changelog: v2.8.0...v2.8.1