The ever-increasing complexity of LLM models often comes at a steep cost: greater computational requirements, increased energy consumption, and slower inference times. Enter model quantization - a powerful technique that can substantially reduce mode...