Best gpu for llama 2 7b reddit. c++ I can achieve about ~50 tokens/s with 7B q4 gguf models.