Workflow
机器翻译
icon
Search documents
全球机器翻译比赛拿下30个语种第1名,腾讯混元翻译模型开源
Sou Hu Cai Jing· 2025-09-02 11:32
Core Insights - Tencent Hunyuan announced the open-source release of its translation model Hunyuan-MT-7B, which has recently won an international translation competition, allowing developers to download and deploy it for free [1][4] - The Hunyuan-MT-7B model supports 33 languages and 5 dialects, showcasing its comprehensive capabilities as a lightweight translation model [1][6] - The model achieved outstanding results in the WMT2025 competition, ranking first in 30 out of 31 languages, demonstrating its superiority over larger models [4][6] Model Features - Hunyuan-MT-7B is characterized by its efficiency, achieving performance that meets or exceeds larger models with only 7 billion parameters [6] - The model's inference speed is significantly faster than that of larger models, allowing it to handle more translation requests under the same hardware conditions [6] - The model can be deployed across various hardware environments, from high-end servers to edge devices, with lower deployment, operational, and maintenance costs [6] Technical Advancements - Tencent Hunyuan has developed a complete training paradigm for translation models, covering pre-training, supervised tuning, and reinforcement learning, which contributes to its industry-leading translation performance [4][6] - The model has been integrated into multiple Tencent services, enhancing user experience across platforms such as Tencent Meeting, WeChat Work, QQ Browser, and more [6] Open Source Commitment - Since its debut in 2023, Tencent Hunyuan has embraced open-source principles, sharing its self-developed technologies and promoting breakthroughs in large model technology [7] - The Hunyuan-MT-7B model is available for experience and download on Tencent Hunyuan's official website, as well as on open-source platforms like Huggingface and GitHub [7]
AI模型终于能翻译“拼多多砍一刀”了
3 6 Ke· 2025-09-02 08:25
Core Insights - Tencent's Hunyuan-MT series translation models, Hunyuan-MT-7B and Hunyuan-MT-Chimera-7B, have been released as open-source, providing advanced machine translation capabilities across 33 languages, including minority languages and dialects [1][7][18] - The models have demonstrated superior performance in various translation benchmarks, outperforming established systems like Google Translate and other models with significantly larger parameters [2][26] - Hunyuan-MT-7B achieved first place in 30 out of 31 language pairs at the WMT 2025 competition, showcasing its effectiveness in both resource-rich and resource-scarce languages [4][22] Model Performance - Hunyuan-MT-7B excels in translating internet slang and gaming terminology, accurately interpreting phrases like "小红薯" as "REDnote" and "砍一刀" as Pinduoduo's pricing mechanism, while Google Translate fails to capture the context [8][10] - The model's understanding of cultural nuances and idiomatic expressions allows it to produce translations that are more natural and contextually appropriate compared to traditional systems [12][13] - In professional translation tests, Hunyuan-MT-7B showed strong capabilities in translating specialized terms, although it still faced challenges in maintaining fluency in sentence structure [15][18] Technical Advancements - The models utilize a "weak-to-strong" reinforcement learning approach, integrating multiple candidate translations to enhance output quality beyond single candidates [5][24] - Hunyuan-MT-7B has been optimized for performance, achieving a 30% improvement in inference speed through FP8 quantization [7][26] - The training data for these models includes a vast dataset of 1.3 trillion tokens from over 112 languages, ensuring a diverse and high-quality training foundation [19][21] Future Implications - The advancements in machine translation technology, particularly through the use of generative AI, are expected to significantly enhance cross-border business operations for companies like Tencent, ByteDance, and Alibaba [28] - The ongoing development of translation models indicates a trend towards more sophisticated and efficient solutions in the field of computational linguistics, potentially leading to broader applications in various industries [28]
同花顺:上半年净利润同比增长38.29% 拟10派1元
Ge Long Hui A P P· 2025-08-22 12:13
Core Insights - The company Tonghuashun (300033.SZ) reported a revenue of 1.779 billion yuan for the first half of 2025, representing a year-on-year growth of 28.07% [1] - The net profit attributable to shareholders reached 502 million yuan, marking a year-on-year increase of 38.29% [1] - The company plans to distribute a cash dividend of 1 yuan (including tax) for every 10 shares to all shareholders [1] Revenue Drivers - The increase in revenue is attributed to the recovery of the capital market, which led to higher user activity on the company's website and app, resulting in increased income from advertising and internet promotion services [1] - There was a notable rise in demand for financial information services from investors, contributing to the growth in value-added telecommunications service revenue [1] Technological Advancements - During the reporting period, the company made significant breakthroughs in various technologies, including large models, intelligent voice, natural language processing, machine translation, and graphics and images [1]
理想VLA实质是强化学习占主导的持续预测下一个action token
理想TOP2· 2025-08-11 09:35
Core Viewpoints - The article presents four logical chains regarding the understanding of "predict the next token," which reflects different perceptions of the potential and essence of LLMs or AI [1] - Those who believe that predicting the next token is more than just probability distributions are more likely to recognize the significant potential of LLMs and AI [1] - A deeper consideration of AI and ideals can lead to an underestimation of the value of what ideals accomplish [1] - The ideal VLA essentially focuses on reinforcement learning dominating the continuous prediction of the next action token, similar to OpenAI's O1O3, with auxiliary driving being more suitable for reinforcement learning than chatbots [1] Summary by Sections Introduction - The article emphasizes the importance of Ilya's viewpoints, highlighting his significant contributions to the AI field over the past decade [2][3] - Ilya's background includes pivotal roles in major AI advancements, such as the development of AlexNet, AlphaGo, and TensorFlow [3] Q&A Insights - Ilya challenges the notion that next token prediction cannot surpass human performance, suggesting that a sufficiently advanced neural network could extrapolate behaviors of an idealized person [4][5] - He argues that predicting the next token well involves understanding the underlying reality that leads to the creation of that token, which goes beyond mere statistics [6][7] Ideal VLA and Reinforcement Learning - The ideal VLA operates by continuously predicting the next action token based on sensor information, indicating a real understanding of the physical world rather than just statistical probabilities [10] - Ilya posits that the reasoning process in the ideal VLA can be seen as a form of consciousness, differing from human consciousness in significant ways [11] Comparisons and Controversial Points - The article asserts that auxiliary driving is more suited for reinforcement learning compared to chatbots due to clearer reward functions [12][13] - It highlights the fundamental differences in the skills required for developing AI software versus hardware, emphasizing the unique challenges and innovations in AI software development [13]