Enabling High-Sparsity Foundational Llama Models with Efficient Pretraining and Deployment.
Abhinav AgarwallaAbhay GuptaAlexandre MarquesShubhra PanditMichael GoinEldar KurticKevin LeongTuan NguyenMahmoud SalemDan AlistarhSean LieMark KurtzPublished in: CoRR (2024)