Core Insights - DeepSeek Company released its new model architecture DeepSeek-V3.2, which has garnered significant industry attention [1] - Cambricon announced its adaptation to DeepSeek-V3.2 and open-sourced the large model inference engine vLLM [1][2] - The DeepSeek-V3.2-Exp model introduces DeepSeek Sparse Attention, optimizing training and inference efficiency for long texts [1] Company Developments - DeepSeek-V3.2-Exp is an experimental version built on V3.1-Terminus, focusing on sparse attention mechanisms [1] - The official DeepSeek applications and APIs have been updated to V3.2-Exp, with significant price reductions to encourage user testing and feedback [1] - Cambricon's adaptation to DeepSeek-V3.2-Exp indicates prior collaboration and technical communication between the two companies [2] Industry Trends - The rapid adaptation of Cambricon to the new model reflects a significant signal of deep collaboration among top Chinese tech companies [2] - The large model has a substantial size of 671GB, requiring approximately 8-10 hours to download under ideal bandwidth conditions [2] - The collaboration between leading companies in the AI chip and model sectors is seen as a strong example of innovation and cooperation in China's tech industry [2]
强强联手!深度求索、寒武纪同步发布DeepSeek-V3.2模型架构和基于vLLM的模型适配源代码