How does DeepSeek-V2 reduce training costs?

Question

Answers ( 1 )

    0
    2025-03-28T02:38:37+00:00

    DeepSeek-V2 reduces training costs by employing a Mixture of Experts (MoE) architecture and sparse computation. These techniques allow the model to activate only a subset of experts during training, thereby reducing the overall computational load and saving 42.5% of the training cost compared to its predecessor, DeepSeek 67B.

Leave an answer