📚 DeepSeek-V2.5 Released by DeepSeek-AI: A Cutting-Edge 238B Parameter Model Featuring Mixture of Experts (MoE) with 160 Experts, Advanced Chat, Coding, and 128k Context Length Capabilities
Nachrichtenbereich: 🔧 AI Nachrichten
🔗 Quelle: marktechpost.com
DeepSeek-AI has released DeepSeek-V2.5, a powerful Mixture of Experts (MOE) model with 238 billion parameters, featuring 160 experts and 16 billion active parameters for optimized performance. The model excels in chat and coding tasks, with cutting-edge capabilities such as function calls, JSON output generation, and Fill-in-the-Middle (FIM) completion. With an impressive 128k context length, DeepSeek-V2.5 […]
The post DeepSeek-V2.5 Released by DeepSeek-AI: A Cutting-Edge 238B Parameter Model Featuring Mixture of Experts (MoE) with 160 Experts, Advanced Chat, Coding, and 128k Context Length Capabilities appeared first on MarkTechPost.
...