Releases: decoderesearch/SAELens
v1.7.0
v1.7.0 (2024-05-08)
Feature
-
feat: Add torch compile (#129)
-
Surface # of eval batches and # of eval sequences
-
fix formatting
-
config changes
-
add compilation to lm_runner.py
-
remove accidental print statement
-
formatting fix (
5c41336) -
feat: Change eval batch size (#128)
-
Surface # of eval batches and # of eval sequences
-
fix formatting
-
fix print statement accidentally left in (
758a50b)
v1.6.1
v1.6.0
v1.5.0
v1.5.0 (2024-05-07)
Feature
Unknown
-
major: How we train saes replication (#123)
-
l1 scheduler, clip grad norm
-
add provisional ability to normalize activations
-
notebook
-
change heuristic norm init to constant, report b_e and W_dec norms (fix tests later)
-
fix mse calculation
-
add benchmark test
-
update heuristic init to 0.1
-
make tests pass device issue
-
continue rebase
-
use better args in benchmark
-
remove stack in get activations
-
broken! improve CA runner
-
get cache activation runner working and add some tests
-
add training steps to path
-
avoid ghost grad tensor casting
-
enable download of full dataset if desired
-
add benchmark for cache activation runner
-
add updated tutorial
-
format
Co-authored-by: Johnny Lin <[email protected]> (5f46329)
v1.4.0
v1.4.0 (2024-05-05)
Feature
-
feat: Store state to allow resuming a run (#106)
-
first pass of saving
-
added runner resume code
-
added auto detect most recent checkpoint code
-
make linter happy (and one small bug)
-
blak code formatting
-
isort
-
help pyright
-
black reformatting:
-
activations store flake
-
pyright typing
-
black code formatting
-
added test for saving and loading
-
bigger training set
-
black code
-
move to pickle
-
use pickle because safetensors doesn't support all the stuff needed for optimizer and scheduler state
-
added resume test
-
added wandb_id for resuming
-
use wandb id for checkpoint
-
moved loaded to device and minor fixes to resuming
Co-authored-by: David Chanin <[email protected]> (4d12e7a)
Unknown
-
Fix: sparsity norm calculated at incorrect dimension. (#119)
-
Fix: sparsity norm calculated at incorrect dimension.
For L1 this does not effect anything as essentially it's calculating the abs() and average everything. For L2 this is problematic as L2 involves sum and sqrt. Unexpected behaviors occur when x is of shape (batch, sen_length, hidden_dim).
-
Added tests.
-
Changed sparsity calculation to handle 3d inputs. (
ce95fb2)
v1.3.0
v1.3.0 (2024-05-03)
Feature
-
feat: add activation bins for neuronpedia outputs, and allow customizing quantiles (#113) (
05d650d) -
feat: Update for Neuropedia auto-interp (#112)
-
cleanup Neuronpedia autointerp code
-
Fix logic bug with OpenAI key
Co-authored-by: Joseph Bloom <[email protected]> (033283d)
-
feat: SparseAutoencoder.from_pretrained() similar to transformer lens (#111)
-
add partial work so David can continue
-
feat: adding a SparseAutoencoder.from_pretrained() function
Co-authored-by: jbloomaus <[email protected]> (617d416)
Fix
v1.2.0
v1.1.0
v1.0.0
v1.0.0 (2024-04-27)
Breaking
- chore: empty commit to bump release
BREAKING CHANGE: v1 release (2615a3e)
Chore
-
chore: fix outdated lr_scheduler_name in docs (#109)
-
chore: fix outdated lr_scheduler_name in docs
-
add tutorial hparams (
7cba332)
Unknown
- BREAKING CHANGE: 1.0.0 release
BREAKING CHANGE: 1.0.0 release (c23098f)