What inference options are available for DeepSeek-Coder?
Question
Lost your password? Please enter your email address. You will receive a link and will create a new password via email.
Lorem ipsum dolor sit amet, consectetur adipiscing elit.Morbi adipiscing gravdio, sit amet suscipit risus ultrices eu.Fusce viverra neque at purus laoreet consequa.Vivamus vulputate posuere nisl quis consequat.
Answers ( 1 )
DeepSeek-Coder supports inference via:
- **vLLM**: For efficient text and chat completion.
- **GGUF Quantization**: Compatible with `llama.cpp`.
- **GPTQ**: Supported via `exllamav2` with HuggingFace Tokenizer integration.
- **Direct API Calls**: Using the provided model endpoints.