Skip to content

csalt-research/DeFT-X

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DeFT-X: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer

arXiv Paper

Installation

conda create -n deftx python=3.10
conda activate deftx
conda install pytorch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 pytorch-cuda=12.1 -c pytorch -c nvidia

Download and install DeFT-X:

git clone https://github.com/csalt-research/DeFT-X.git
cd DeFT-X
pip install -e .

Training SFTs

DeFTArguments that may be useful:

  • rank_variant - choice of rank variant to use: uniform or variance.
  • rank - controls the rank value when rank_variant == uniform.
  • variance_threshold - variance threshold to decides the rank value when rank_variant == variance.
  • keep_raio - controls the percentage of parameters to be kept while pruning the high-frequency components.
  • ft_params_num/ft_params_proportion - controls the number/proportion of the maskable params that will be fine-tuned.
  • full_ft_max_steps_per_iteration/full_ft_max_epochs_per_iteration - controls the maximum number of steps/epochs durin full-finetuning. Both can be set.
  • sparse_ft_max_steps_per_iteration/sparse_ft_max_epochs_per_iteration - controls the maximum number of steps/epochs during sparse fine-tuning. Both can be set.
  • full_ft_min_steps_per_iteration/sparse_ft_min_steps_per_iteration - controls the minimum number of steps during fine-tuning. Takes effect if a max number of epochs is set which amounts to a lesser number of steps.

Example Scripts

Examples of DeFT-X training and evaluation are provided in examples/.

Citation

@misc{simon2025deftxdenoisedsparsefinetuning,
      title={DeFTX: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer}, 
      author={Sona Elza Simon and Preethi Jyothi},
      year={2025},
      eprint={2505.15090},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2505.15090}, 
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages