Answers ( 2 )

    0
    2025-04-01T05:11:51+00:00

    DeepSeek-Coder is an open-source series of code language models designed for code generation and completion. It is trained on 2 trillion tokens, with 87% being code and 13% natural language (English and Chinese). The models range from 1B to 33B parameters and support 67 programming languages.

    0
    2025-04-01T05:12:05+00:00

    - **Training Data**: 2 trillion tokens (87% code, 13% natural language).
    - **Model Sizes**: 1B, 5.7B, 6.7B, and 33B parameters.
    - **Performance**: State-of-the-art results on benchmarks like HumanEval, MultiPL-E, MBPP, DS-1000, and APPS.
    - **Functionality**: Code completion, insertion, chat model inference, and repository-level code generation.
    - **Window Size**: 16K tokens for project-level context.

Leave an answer