Sparse Finetuning for Inference Acceleration of Large Language Models
Paper
• 2310.06927 • Published
• 15
Explore our breakthrough in sparse fine-tuning LLMs! Our novel method maintains downstream accuracy even with >70% sparsity.
Solve math problems with chat-based guidance