nateraw / llama-2-70b-chat-awq

llama-2-70b-chat quantized with AWQ and served with vLLM

  • Public
  • 77 runs
  • GitHub
  • Paper
  • License
  1. Author
    @nateraw
    Version
    22.04

    9bf9d123

    Latest
  2. Author
    @nateraw
    Version
    22.04