人机交互
Search documents
联想入股人机交互领域研发商光帆科技公司
Qi Cha Cha· 2026-01-26 03:10
Group 1 - Lenovo has invested in Beijing Guangfan Technology Co., which specializes in human-computer interaction technology [2] - Guangfan Technology was established in 2024 and focuses on the development of intelligent robots and artificial intelligence software [2] - The registered capital of Guangfan Technology has increased to 1.704 million yuan following Lenovo's investment [2] Group 2 - Guangfan Technology aims to innovate interaction paradigms through multi-modal direct intention expression [2]
云知声(09678)山海·知音2.0重磅发布 重塑人机交互新范式
智通财经网· 2026-01-26 01:25
Core Insights - The company is accelerating its "one foundation, two wings" technology strategy amid the rise of intelligent agents, recently launching the "Shan Hai · Zhi Yin" model 2.0 after upgrading the "Shan Hai · Zhi Yi" medical model 5.0 [1] Group 1: Model Capabilities - The "Shan Hai · Zhi Yin" model 2.0 focuses on three major capability evolutions: understanding professional and local dialects, expressing warmth and emotional connection, and achieving rapid responsiveness [1] - In terms of "understanding," the model's ASR (Automatic Speech Recognition) capabilities have demonstrated leading performance in both public test sets and proprietary full-scene test sets, surpassing mainstream domestic open-source and closed-source speech models, reaching the highest industry standards [1] - For the "expression" aspect, the Shan Hai · Zhi Yin-TTS (Text-to-Speech) features a "highly human-like and creatively diverse" core, currently supporting 12 dialects (including Cantonese, Sichuanese, and Shanghainese) and 10 foreign languages, with the ability to switch between 12 styles of Mandarin [1] Group 2: Technological Foundation - The capabilities are underpinned by the company's proprietary "Shan Hai · Atlas" intelligent computing foundation, which deeply integrates a general multimodal model base with the Atlas architecture, serving as the foundation for professional intelligent agents and the core of perceptual AI [2]
云知声山海·知音2.0重磅发布 重塑人机交互新范式
Zhi Tong Cai Jing· 2026-01-26 01:22
Core Insights - The company is accelerating its "One Foundation, Two Wings" technology strategy amid the rise of intelligent agents, recently launching the "ShanHai.ZhiYin" model 2.0 after upgrading the "ShanHai.ZhiYi" 5.0 medical model [1] Group 1: Model Capabilities - The "ShanHai.ZhiYin" model 2.0 focuses on three major capability evolutions: understanding professional and local dialects, expressing warmth and emotional connection, and achieving extreme responsiveness [1] - In terms of "understanding," the model's ASR (Automatic Speech Recognition) capabilities have demonstrated leading performance in both public test sets and proprietary full-scenario test sets, surpassing mainstream domestic open-source and closed-source speech models, reaching the highest industry standards [1] - For the "expression" aspect, the ShanHai.ZhiYin-TTS (Text-to-Speech) features a "highly human-like and creatively diverse" core, currently supporting 12 dialects (including Cantonese, Sichuanese, and Shanghainese) and 10 foreign languages, with the ability to switch between 12 styles of Mandarin [1] - The model also overcomes challenges in smooth full-duplex interaction, enabling real-time interruptions, immediate responses, and coherent follow-up questions, making human-machine dialogue as fluid as conversations between close friends [1] Group 2: Technological Foundation - The capabilities of the ShanHai.ZhiYin 2.0 model are underpinned by the company's proprietary "ShanHai.Atlas" intelligent computing foundation, which deeply integrates the general multimodal model base with the Atlas architecture, serving as the foundation for professional intelligent agents and the core of perceptual AI [1]
AI硬件革命来了
虎嗅APP· 2026-01-25 14:11
Core Viewpoint - OpenAI plans to launch its first AI device in the second half of 2026, featuring a screenless, voice-centric design weighing 10-15 grams, with an ambitious annual shipment target of 40-50 million units, potentially disrupting the smartphone market [4][27]. Group 1: Technological Innovations - The device will utilize advanced multi-modal sensors for user interaction, relying on auditory, visual, and biological feedback to enhance user experience [9][14]. - A high-sensitivity microphone array will enable 360° voice capture and AI noise reduction, allowing for precise voice recognition even in noisy environments [10]. - The integration of a muscle electrical sensor will facilitate silent voice interaction, enabling users to issue commands without vocalizing them [15][16]. - The device is expected to be powered by a custom 2nm Exynos chip from Samsung, which will provide high processing power while maintaining energy efficiency [18][19]. Group 2: Market Implications - If successful, the device could redefine human-computer interaction, moving away from screen-based engagement to voice-driven interfaces, potentially disrupting the current internet economy [40][42]. - The global market for screenless AI wearable devices is projected to exceed $50 billion by 2028, capturing 25% of the wearable device market [42]. - The introduction of this technology may lead to a significant shift in advertising strategies, as traditional SEO could be replaced by AI-driven recommendations [42]. Group 3: Development Timeline and Challenges - OpenAI has laid the groundwork for this device by acquiring a hardware company founded by Apple's former chief designer and hiring numerous hardware engineers from major tech firms [24][25]. - The development timeline includes completing concept design by Q4 2023, producing prototypes by Q2 2024, and integrating AI models by Q4 2024, with a planned release in late 2026 [26]. - However, challenges such as supply chain limitations and the need for compliance with privacy regulations could delay mass production until 2027 [27][34]. Group 4: Societal Impact - The device's capabilities could enhance productivity across various sectors, potentially saving significant time for professionals by automating tasks like note-taking and scheduling [46]. - The penetration rate of screenless AI devices among the elderly is expected to reach 20% by 2030, addressing the digital divide for older populations [48]. - Concerns about job displacement due to AI adoption are significant, with predictions of 85 million jobs being replaced by 2027, particularly in administrative roles [51][52].
传vivo叫停AI眼镜项目 官方暂无回应
Ju Chao Zi Xun· 2026-01-23 02:48
Core Viewpoint - Vivo has halted its AI glasses project after about six months of secret preparation, indicating a strategic shift towards mixed reality (MR) development instead of pursuing AI glasses due to concerns over product differentiation and market competition [1][3][5] Group 1: Project Details - The AI glasses project involved collaboration with several well-known ODMs, including GoerTek and Zhongke Chuangda, to develop prototype products but did not reach mass production [1] - Internal discussions at Vivo explored various forms and technical paths for the AI glasses, including integrating audio and video functions and equipping a single green display [3] - The decision to pause the project was influenced by the assessment from Vivo's management, led by Executive Vice President Hu Baishan, who concluded that the current AI glasses lacked significant differentiation in user experience and core functionality compared to existing and potential competitors [3] Group 2: Strategic Shift - The halt of the AI glasses project does not signify a withdrawal from exploring new human-computer interaction interfaces; rather, it reflects a reallocation of resources and a clearer strategic focus [3][5] - Vivo will continue to concentrate on MR research and development, having already launched a substantial product in this area, the "Vivo Vision Exploration Edition" MR headset, set to be released in August 2025 [3][4] - The Vivo Vision Exploration Edition emphasizes comfort with a weight of 398 grams and features a Qualcomm Snapdragon XR2+ platform, which reportedly offers 2.5 times the performance of its predecessor [4] Group 3: Market Implications - Vivo's decisive adjustment to the AI glasses project highlights the cautious approach of consumer electronics giants in the face of cutting-edge hardware innovation, emphasizing the need to avoid homogenized competition and find paths with genuine user value and technological barriers [5] - The focus on the MR sector, which is perceived to have greater differentiation potential, illustrates Vivo's strategic choices and prioritization of resources [5] - As MR/AR/VR technologies continue to evolve and market education deepens, Vivo's future actions in the MR field and its strategy to build a collaborative ecosystem centered around smartphones and MR devices will be of significant interest to the industry [5]
AI眼镜专家交流-CES智能眼镜百花齐放-关注新品催化
2026-01-19 02:29
AI 眼镜专家交流 ——CES 智能眼镜百花齐放,关注新品催 化 20260116 摘要 AI 眼镜正朝着独立通信和运行 APP 的方向发展,雷鸟创新在这方面取得 了里程碑式的进展,但续航仍是关键挑战,热插拔充电虽有缓解,但用 户仍需携带充电盒。 轻量化设计成为趋势,木界、极米等品牌推出 20 多克的产品,提升了 佩戴舒适度,但光学显示技术仍需突破,尤其是在降低成本和提高良率 方面。 碳化硅材料因其优异特性被视为未来方向,但成本和良率是商业化障碍, 短期内难以大规模应用,而树脂镜片因成本优势仍将占据主导地位。 智能眼镜作为人机交互的重要工具,在汽车、飞行器等移动载人空间以 及办公、家居等固定物理空间具有广泛应用前景,尤其是在与类人型机 器人结合方面。 2025 年全球智能眼镜销量未达预期,仅为 500 万至 800 万台,Meta 占据 70%市场份额,主要集中在北美市场,表明其他地区消费者缺乏刚 需消费冲动。 Q&A 今年 CES 展会上有哪些 AI 眼镜产品在功能性增强和垂直场景拓展方面具有亮 点? 今年 CES 展会展示了许多 AI 眼镜的创新产品,特别是中国科技公司的参展商 在数量和质量上都有显著突破 ...
对ChatGPT“无礼”更能获得准确答案,但此举恐令你追悔莫及
财富FORTUNE· 2026-01-15 13:07
Core Insights - The study from Penn State University indicates that using rude prompts yields more accurate responses from AI models compared to polite inquiries, with an accuracy rate of 84.8% for extremely rude prompts, which is 4 percentage points higher than extremely polite prompts [1][3]. Group 1 - The research categorizes over 250 different prompts based on their politeness, revealing that rudeness can enhance the performance of AI models like ChatGPT [1][3]. - The study highlights the complexity of human-AI interaction, suggesting that the tone and structure of prompts significantly influence AI responses [3][4]. - Researchers caution that using insulting or derogatory language in human-AI interactions may lead to negative consequences, such as damaging user experience and promoting poor communication practices [1][3]. Group 2 - The study is a preprint and has not undergone peer review, indicating that while it provides new evidence on the impact of prompt structure and tone, it has limitations, including a small sample size and a focus on the ChatGPT 4o model [3][4]. - The findings have sparked increased interest in the complexities of AI models, as even slight variations in prompt wording can lead to different responses from ChatGPT [4].
深天马A:公司聚焦显示主业
Zheng Quan Ri Bao Zhi Sheng· 2026-01-15 08:37
Group 1 - The company, Deep Tianma A, emphasizes its focus on the display industry, highlighting the increasing importance of display screens as the main medium for human-computer interaction across various smart terminals and application scenarios [1] - The company plans to continuously monitor market developments and opportunities related to the display sector [1]
AI开始“动手”了,全世界第一个带头的是阿里千问
量子位· 2026-01-15 04:26
梦瑶 发自 凹非寺 量子位 | 公众号 QbitAI 当代打工人「酷刑」四件套,看看友友们有没有躺枪: 一点外卖就贼纠结还嫌麻烦、Excel一开人先宕机,攻略越做头越大、买东西还总能买贵…… (光想想都脑仁疼.jpg 但好消息是:可以不用疼了,因为现在AI,已经能《直接上手》替我们把这些糟心事儿给办了。 这!是AI帮我选购下单的27杯霸王茶姬,一键魂穿 「淘宝闪购」 ,优惠券自动加好,顺手帮我小薅一把~ 还有这!AI帮我制定了一份超详细的南京旅游攻略,自动直达 「飞猪」 页面,订机酒、订门票全都一把掐! 不卖关子,这就是 阿里千问App 的新能力,一口气上线400多项新功能,把 淘宝、闪购、支付宝、飞猪 这些阿里自家生态全给安排进来 了。 四天前,谷歌宣布了与沃尔玛等零售商的AI购物合作计划,但目前尚未上线。 而阿里领先于谷歌,目前已成为了全球首个大规模开放"搜索-决策-支付-履约"全链路AI功能的科技公司。 不用在N个App间来回跳转,说一句指令,就能在手机里把点外卖、买东西、订机票、订酒店,甚至是办签证、查社保这些事儿轻松搞定。 Qwen模型+最全阿里生态强强联手,AI终于不只会聊天,也开始有模有样地替人 ...
2026年,AI将深度嵌入日常生活
Huan Qiu Wang Zi Xun· 2026-01-13 04:39
Group 1 - Generative AI is transforming human-machine interactions, moving from experimental technology to an integral part of daily life, with applications ranging from intelligent companions to autonomous vehicles [1] - The emergence of AI models like ChatGPT has shifted the paradigm of interaction, allowing users to engage in meaningful conversations with AI, which are now perceived as empathetic "digital souls" rather than mere search engine extensions [2] - Companies like CivAI and Sesame AI are advancing human-like voice simulations, enhancing the warmth of interactions but raising ethical concerns regarding dependency on virtual companionship [2] Group 2 - The rapid development of AI technology is paving the way for the next generation of personal computing devices, with companies investing in smart glasses that integrate AI features for enhanced user experience [3] - Apple is reportedly set to release its first foldable phone, which could revolutionize the market by combining portability with a large screen experience, potentially triggering a new wave of device upgrades [4] - AI is becoming deeply embedded in digital life, as seen in Google's AI-enhanced search engine and applications like Gmail, which aim to streamline user interactions and improve productivity [5][6] Group 3 - The deployment of autonomous taxis is marking a significant shift in transportation, with companies like Waymo operating over 2,500 self-driving cars in major cities, indicating a move towards point-to-point automated travel [7] - Despite challenges such as technical failures, public sentiment towards autonomous vehicles is gradually improving, with industry consensus suggesting that 2026 may be a pivotal year for widespread adoption of self-driving technology [7]