An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. (開放原始碼的專家混合模型程式碼語言模型,其效能可與 GPT4-Turbo 在程式碼特定任務中相媲美。)

16b 236b

536.5K 5 個月前

讀我檔案

DeepSeek-Coder-V2 is an open-source Mixture-of-Experts (MoE) code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. DeepSeek-Coder-V2 is further pre-trained from DeepSeek-Coder-V2-Base with 6 trillion tokens sourced from a high-quality and multi-source corpus. (DeepSeek-Coder-V2 是一個開放原始碼的專家混合模型 (MoE) 程式碼語言模型,其效能可與 GPT4-Turbo 在程式碼特定任務中相媲美。DeepSeek-Coder-V2 是從 DeepSeek-Coder-V2-Base 進一步預訓練而來,使用了來自高品質和多來源語料庫的 6 兆個 tokens。)

參考資料

Hugging Face