Scaling Sparse Fine-Tuning to Large Language Models

Published in arxiv preprint, 2024

We scale sparse fine-tuning methods for 7b and 13b Llama 2, such that the memory requirements scale linearly with the number of paramerters updated by the fine-tuning.

Full Paper