Ansh MittalREVOLUTIONIZING DEEP LEARNING WITH SYSML: The Power of 2-Bit Quantized Neural NetworksThis post is the article for the first SysML paper I read this year. I will be posting regularly for all my read papers. TL;DR: In the deep…Jun 19, 2023Jun 19, 2023
Benjamin MarieFrom 16-bit to 2-bit: Finding the Best Trade-off Between Memory-Efficiency and AccuracyMistral 7B and Llama 2 under pressureFeb 6, 20241Feb 6, 20241
InLevel Up CodingbyYeyu HuangThe 2-bit Quantization is Insane! See How to Run Mixtral-8x7B on Free-tier Colab.A Quick Tutorial for AQLM-2-bit Quantization and its ImplementationFeb 20, 20242Feb 20, 20242