cancel
Showing results for 
Search instead for 
Did you mean: 
m_koch_unify
New Contributor II
since ‎07-20-2023
‎07-21-2023

User Stats

  • 2 Posts
  • 0 Solutions
  • 1 Kudos given
  • 0 Kudos received

User Activity

Hi all,I was following the hugging face model https://huggingface.co/TheBloke/Llama-2-70B-chat-GPTQ, which points to use Exllama (https://github.com/turboderp/exllama/), which has 4 bit quantization.Running on a A10-Single-GPU-64GB,I've cloned the Ex...
Kudos given to