Qwen 2.5 (32B)
Search documents
老外傻眼,明用英文提问,DeepSeek依然坚持中文思考
3 6 Ke· 2025-12-03 09:14
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which show significant improvements in reasoning capabilities, with DeepSeek-V3.2 competing directly with GPT-5 and Speciale performing comparably to Gemini-3.0-Pro [1] - There is a notable phenomenon where even when queries are made in English, the model sometimes reverts to using Chinese during its reasoning process, leading to confusion among overseas users [3][5] - The prevalent belief is that Chinese characters have a higher information density, allowing for more efficient expression of the same textual meaning compared to English [5][9] Model Performance and Efficiency - Research indicates that using non-English languages for reasoning can lead to a 20-40% reduction in token consumption without sacrificing accuracy, with DeepSeek R1 showing token reductions ranging from 14.1% (Russian) to 29.9% (Spanish) [9] - A study titled "EfficientXLang" supports the idea that reasoning in non-English languages can enhance token efficiency, which translates to lower reasoning costs and reduced computational resource requirements [6][9] - Another study, "One ruler to measure them all," reveals that English is not the best-performing language for long-context tasks, ranking sixth among 26 languages, with Polish taking the top spot [10][15] Language and Training Data - The observation that Chinese is frequently used in reasoning by models trained on substantial Chinese datasets is considered normal, as seen in the case of the AI programming tool Cursor's new version [17] - The phenomenon of models like OpenAI's o1-pro occasionally using Chinese during reasoning is attributed to the higher proportion of English data in their training, which raises questions about the language selection process in large models [20] - The increasing richness of Chinese training data suggests that models may eventually exhibit more characteristics associated with Chinese language processing [25]
老外傻眼!明用英文提问,DeepSeek依然坚持中文思考
机器之心· 2025-12-03 08:30
Core Insights - DeepSeek has launched two new models, DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, which show significant improvements in reasoning capabilities, with the former being comparable to GPT-5 and the latter performing similarly to Gemini-3.0-Pro [1][4] - There is a notable phenomenon where DeepSeek switches to Chinese during reasoning, even when queries are made in English, leading to discussions about the efficiency of Chinese in processing information [4][6] Group 1: Model Performance - The new models exhibit enhanced reasoning speed, attracting interest from overseas researchers [1] - The comment section reflects a consensus that Chinese characters have a higher information density, requiring fewer characters to express the same meaning compared to English [4][6] Group 2: Cross-Lingual Reasoning - Research indicates that using non-English languages for reasoning can lead to better performance and reduced token consumption, as shown in the paper "EfficientXLang" [7][8] - The study found that reasoning in non-English languages can achieve a token reduction of 20-40% without sacrificing accuracy, with DeepSeek R1 showing reductions from 14.1% (Russian) to 29.9% (Spanish) [11] Group 3: Language Efficiency - Although Chinese can save reasoning token costs compared to English, it is not the most efficient language; Polish ranks highest in long-context tasks [12][14] - The performance of models varies significantly based on the language used for instructions, with English not being the top performer in long-context tasks [14][18] Group 4: Training Data Influence - The prevalence of Chinese training data in domestic models explains the tendency for these models to think in Chinese [20][21] - The phenomenon of models like OpenAI's o1-pro occasionally using Chinese during reasoning raises questions about the influence of training data composition [24][25]