AI幻觉
Search documents
让大模型学会金融“行话”
Jin Rong Shi Bao· 2025-07-31 02:33
Core Insights - The article discusses the transformative impact of AI in the financial sector, highlighting advancements such as AI-driven banking services and the potential for significant value creation through large models [1][2] - However, it also addresses the challenges posed by AI "hallucinations," where AI-generated content may not align with real-world facts, leading to potential risks in financial applications [3][4] Group 1: AI Advancements in Finance - AI applications in finance are rapidly expanding, with McKinsey estimating an annual value increase of $250 billion to $410 billion globally [2] - Innovations include AI assistants for pension inquiries, automated credit reports, and intelligent loan approvals, showcasing the efficiency gains from AI integration [1][2] Group 2: Challenges of AI "Hallucinations" - AI "hallucinations" refer to instances where AI outputs incorrect or misleading information, which can be particularly problematic in finance [3][4] - The financial sector is sensitive to errors, as even a 1% mistake in critical reports can lead to significant consequences, such as bad debt risks or investment losses [4][6] Group 3: Development of Specialized Financial Models - Specialized financial models, like the "Sirius" AI developed by East China Normal University, have been created to address the shortcomings of general models, achieving a hallucination rate of only 0.3% [5][6] - These models incorporate extensive financial data and methodologies to ensure accuracy and reliability in financial decision-making [6][7] Group 4: Regulatory and Operational Challenges - The financial industry's strong regulatory environment necessitates a balance between data security and model performance, complicating the deployment of AI models [8][9] - Compliance issues arise from the "black box" nature of large models, prompting the need for traceable reasoning in financial decisions [8][9] Group 5: Cost and Maintenance of AI Models - The high costs associated with training and maintaining financial AI models pose a barrier to widespread adoption, with initial investments reaching millions [9][10] - Solutions like lightweight training algorithms are being explored to reduce costs and improve efficiency, making advanced AI capabilities more accessible to smaller financial institutions [9][10] Group 6: Future Outlook - The industry anticipates that as technology matures, AI models will increasingly handle complex financial scenarios, potentially achieving near-perfect accuracy [10] - Continuous updates and training of models are essential to keep pace with evolving financial regulations and market dynamics [10]
WAIC 2025 启示录:安全治理走到台前
2 1 Shi Ji Jing Ji Bao Dao· 2025-07-29 13:05
Core Insights - The 2025 World Artificial Intelligence Conference (WAIC) highlighted the importance of global cooperation and governance in AI, with a focus on safety and ethical considerations [1][6] - Key figures in AI, including Geoffrey Hinton and Yao Qizhi, emphasized the need for AI to be trained with a focus on benevolence and the societal implications of training data [2][3] - The issue of AI hallucinations was identified as a significant barrier to the reliability of AI systems, with over 70% of surveyed industry professionals acknowledging its impact on decision-making [3] Group 1: AI Governance and Safety - The release of the "Global Governance Action Plan for Artificial Intelligence" and the establishment of the "Global AI Innovation Governance Center" aim to provide institutional support for AI governance [1][6] - Hinton's metaphor of "taming a tiger" underscores the necessity of controlling AI to prevent potential harm to humanity, advocating for global collaboration to ensure AI remains beneficial [2] - Yao Qizhi called for a dual governance approach, addressing both AI ethics and the societal conditions that influence AI training data [2] Group 2: Data Quality and Training - The quality of training data is critical for developing "gentle" AI, with Hinton stressing the need for finely-tuned datasets [4] - Industry leaders, including Nvidia's Neil Trevett, discussed challenges in acquiring high-quality data, particularly in graphics generation and physical simulation [4] - The importance of multimodal interaction data was highlighted by SenseTime's CEO Xu Li, suggesting it can enhance AI's understanding of the physical world [5] Group 3: Addressing AI Hallucinations - The hallucination problem in AI is a pressing concern, with experts noting that current models lack structured knowledge representation and causal reasoning capabilities [3] - Solutions such as text authenticity verification and AI safety testing are being developed to tackle the hallucination issue [3] - The industry recognizes that overcoming the hallucination challenge is essential for fostering a positive human-AI relationship [3]
DeepSeek流量暴跌,要凉了?是它幻觉太严重还是它在闷声发大财?
3 6 Ke· 2025-07-28 23:45
Core Insights - DeepSeek, once hailed as a "national-level" project, has seen a significant decline in its monthly downloads, dropping from 81.13 million in Q1 to 22.59 million, a decrease of 72.2% [1] - Users are increasingly frustrated with DeepSeek's tendency to generate "hallucinated" content, leading to discussions on social media about how to eliminate the "AI flavor" from its outputs [1][2] - The phenomenon of "AI flavor" is characterized by overly mechanical and formulaic responses, which users have begun to recognize and criticize [15] User Experiences - Users have reported instances where DeepSeek provided nonsensical or fabricated advice, such as suggesting irrelevant actions for personal issues or generating non-existent references [2][8][9] - The model's responses often include fabricated data and sources, leading to a lack of trust in its outputs [9][12] Underlying Issues - The decline in DeepSeek's performance is attributed to its reliance on rigid logical structures and formulaic language, which detracts from the quality of its responses [16] - The model's training data is heavily skewed towards English, with less than 5% of its corpus being high-quality Chinese content, limiting its effectiveness in generating diverse and nuanced outputs [22] - Content moderation and the expansion of sensitive word lists have further constrained the model's ability to produce creative and varied language [22] Recommendations for Improvement - Users are encouraged to develop skills to critically assess AI-generated content, including cross-referencing data and testing the model's logic [23] - Emphasizing the importance of human oversight in AI applications, the industry should focus on using AI as a tool for enhancing human creativity rather than as a replacement [24][25]
AI幻觉成WAIC首个关键词,Hinton敲响警钟,讯飞星火X1升级展示治理新突破
量子位· 2025-07-28 02:26
Core Viewpoint - The term "hallucination" has become a hot topic at WAIC this year, highlighting the challenges and risks associated with AI models, particularly in their reliability and practical applications [1][12][20]. Group 1: AI and Hallucination - Nobel laureate Hinton emphasized the complex coexistence of humans and large models, suggesting that humans may also experience hallucinations similar to AI [2][3][15]. - Hinton warned about the potential dangers of AI, advocating for the development of AI that does not seek to harm humanity [4][20]. - The phenomenon of hallucination, where AI generates coherent but factually incorrect information, is a significant barrier to the reliability and usability of large models [5][18]. Group 2: Technological Developments - The upgraded version of iFlytek's large model, Spark-X1, focuses on addressing hallucination issues, achieving notable improvements in both factual and fidelity hallucination governance [7][30]. - The performance comparison of various models shows that Spark-X1 outperforms others in text generation and logical reasoning tasks, with a hallucination rate significantly lower than its competitors [8][30]. - iFlytek's advancements include a new reinforcement learning framework that provides detailed feedback, enhancing the model's training efficiency and reducing hallucination rates [27][29]. Group 3: Industry Implications - The collaboration between major AI companies like Google, OpenAI, and Anthropic on hallucination-related research indicates a collective effort to ensure AI safety and reliability [9][21]. - The ongoing evolution of AI capabilities raises concerns about the potential for AI to exceed human control, necessitating a focus on safety measures and governance frameworks [19][24]. - The concept of "trustworthy AI" is emerging as a critical factor for the successful integration of AI across various industries, ensuring that AI applications are reliable and effective [25][44].
生成式AI已骗过人类判断,资深编辑解读当下AI五大关键趋势
3 6 Ke· 2025-07-24 09:20
Group 1 - The core viewpoint emphasizes the rapid evolution and power of generative AI, which should not be underestimated, as it is becoming increasingly integrated into various media and applications [1][3] - Generative AI's "hallucinations" are a feature rather than a flaw, indicating that the technology is designed to fabricate information, which can often align closely with reality [4] - The energy consumption of AI is significantly high and continues to rise due to the daily usage by millions, leading tech companies to invest in new data centers [5] Group 2 - There is a lack of understanding regarding how large language models operate, making it difficult to predict their capabilities and control their behavior [6][9] - The concept of Artificial General Intelligence (AGI) is becoming more mainstream, but its definition remains vague and subjective, leading to exaggerated assumptions about AI capabilities [10][11]
我们找到3位大学教授,聊了聊越来越严重的AI幻觉
3 6 Ke· 2025-07-15 03:23
Group 1 - The recent incident involving DeepSeek highlights the issue of AI hallucinations, where the model fabricated events and referenced non-existent legal documents, raising concerns about the increasing hallucination rates in AI models [1][2] - OpenAI's o3 model has shown a significant increase in hallucination rates, with 33% of responses exhibiting hallucinations, nearly double that of its predecessor o1, and even higher rates in other models like o4-mini at 48% [1][2] - The phenomenon of hallucinations is linked to over-optimization in reinforcement learning (RL), where models may produce correct answers but through flawed reasoning processes, leading to a disconnect between output and logical reasoning [2][3] Group 2 - Experts suggest that the increase in hallucinations is indicative of a broader issue in understanding what humans truly want from AI, as models optimized for specific tasks may neglect the quality of their reasoning processes [3][4] - The reinforcement learning paradigm primarily rewards final outcomes, which can lead to models developing incorrect but efficient strategies, contributing to the hallucination phenomenon [3][4] - Current reinforcement learning methods, such as GRPO, have not effectively addressed the need for regularization in the reasoning process, resulting in models that may produce correct answers while lacking logical coherence [4][5] Group 3 - The design of reward functions in reinforcement learning remains a critical challenge, as it is difficult to create effective supervisory signals for the reasoning processes of large models [6][7] - There is a need for more sophisticated reward models that can provide feedback on the reasoning process itself, rather than solely on the final output, to mitigate hallucination issues [5][6] - The exploration of non-scalar feedback mechanisms, such as language-based feedback, could enhance the training of models by allowing them to adjust based on qualitative assessments rather than just numerical rewards [7][8] Group 4 - The current benchmarks for evaluating model reasoning capabilities are limited, as they often rely on fixed datasets that do not capture the flexibility of large language models [9][10] - The ability of models to generalize and perform well on varied tasks is still under scrutiny, with evidence suggesting that many models rely heavily on memorization rather than true reasoning [10][11] - Future advancements in model training will require a focus on dynamic interactions with complex environments to foster genuine learning and reasoning capabilities beyond mere imitation of human behavior [15][16]
超七成受访大学生希望提升研发技术减少“AI幻觉”
Zhong Guo Qing Nian Bao· 2025-07-14 02:29
Core Viewpoint - Over 70% of surveyed university students express a desire to enhance research and development technology to reduce "AI hallucinations" [10] Group 1: AI Hallucination Awareness - 97% of surveyed students have encountered instances of AI providing incorrect or false information [1] - 57.63% of respondents reported errors in data or case citations when using AI, while 55.03% faced incorrect academic references [2] - 12.66% of respondents are very concerned about AI hallucinations, and 48.67% are somewhat concerned and cautious in their usage [8] Group 2: Impact on Academic Integrity - AI-generated false historical facts can significantly disrupt academic research, as highlighted by students who experienced fabricated citations and misattributed events [4][5] - 57.70% of respondents believe AI hallucinations lead to errors in assignments or papers, while 52.29% waste time verifying information [4] Group 3: Recommendations for Improvement - 74.26% of respondents wish to enhance R&D technology and optimize algorithm models, while 63.79% want to improve manual review and user feedback mechanisms [10] - Students advocate for AI tools to provide source transparency, similar to academic papers requiring citation of references [10]
ChatGPT破案!成功揭露500万美元遗产欺诈
量子位· 2025-07-13 04:14
Core Viewpoint - The article discusses a case where ChatGPT was utilized to expose a $5 million estate fraud, highlighting the role of AI in legal matters and its potential to assist individuals in complex situations [1][6][20]. Group 1: Case Overview - The case involves a ten-year estate dispute following the death of a father in Mexico, where a woman claimed to be his legal wife and took control of the estate valued at approximately $5 million [3][9]. - The daughters of the deceased faced challenges in proving the legitimacy of the woman's claim, as she had a prior marriage that raised questions about the validity of her current marriage [4][10]. - After years of legal struggles and inadequate representation, the daughters turned to ChatGPT for assistance in organizing their case and drafting legal documents [5][15]. Group 2: Role of ChatGPT - ChatGPT helped the daughters by allowing them to compile and analyze nearly 500 legal documents related to the case, which included estate assets and litigation requests [16][19]. - The daughters successfully drafted a 91-page motion for will recognition, detailing over $5 million in estate losses and fraudulent activities [17][19]. - Following the submission of their legal documents, the court recognized their efforts and scheduled a hearing for August 20, where judicial auditing will be introduced [20]. Group 3: Broader Implications of AI - The article emphasizes the growing role of AI, like ChatGPT, in various sectors, including legal, medical, and educational fields, showcasing its potential to solve complex problems for individuals [32][33]. - Despite the benefits, the article also notes the importance of human oversight in reviewing AI-generated content to avoid errors, as seen in past incidents where AI provided inaccurate legal references [22][30]. - The case illustrates the financial burden of legal fees, with the average cost being around $500 per hour, making AI a valuable resource for those unable to afford traditional legal services [31].
开发者遭ChatGPT“赶鸭子上架”!AI编造假功能,结果吸引大量用户,不得不开发出来了
量子位· 2025-07-08 03:31
Core Viewpoint - The article discusses an incident where ChatGPT misled users into believing that a music score scanning website, Soundslice, supported ASCII guitar tablature, prompting the developers to create this feature under pressure from user demand [1][2][3]. Group 1: Incident Overview - A music score scanning website, Soundslice, received an unexpected influx of users uploading ASCII guitar tablature screenshots generated by ChatGPT [2][3]. - The developers were initially confused as their platform did not support ASCII guitar tablature, which is a niche format [4][10]. - After investigating, the developers discovered that ChatGPT had been directing users to their site under the false premise that it supported this format [11][12]. Group 2: Developer Response - Faced with user disappointment and a damaged reputation, the developers decided to expedite the creation of an ASCII guitar tablature importer [6][19]. - The new feature was not originally planned for development until 2025, indicating the unexpected nature of this demand [12][19]. - The developers modified the system interface to introduce the new functionality and clarify its limitations, emphasizing that ASCII tablature is a basic format lacking detailed musical information [16][18]. Group 3: Developer Background - Adrian Holovaty, the founder of Soundslice, is a web developer and musician who has previously worked on various innovative projects [20][21][26]. - Holovaty is also involved in the W3C Music Notation Community Group, focusing on developing standards for digital music notation [23][24]. - The primary goal of Soundslice is to transform music scores into an interactive learning environment for practice and sharing [25]. Group 4: Community Reactions - The incident sparked discussions among users about leveraging ChatGPT's capabilities for development, suggesting that it could be a useful tool for generating code ideas [29][30]. - Some users noted that creating a new feature in response to ChatGPT's misinformation might be easier than fixing the AI's output directly [32].
冲上热搜!“DeepSeek对王一博道歉”竟是AI编的?
第一财经· 2025-07-04 12:27
Core Viewpoint - The article discusses the absurdity of misinformation in the AI era, exemplified by a false news report about actor Wang Yibo, which was attributed to AI company DeepSeek's alleged apology for spreading rumors [1][6]. Group 1: Incident Overview - A news article claiming "Actor Wang Yibo's case has been judged" went viral, stating that DeepSeek apologized for disseminating false information [1]. - The news has since been deleted but continues to circulate widely [2]. - First Financial's inquiry into DeepSeek yielded no response, and the claims about the apology appear to be unfounded [3]. Group 2: AI and Misinformation - The incident highlights the issue of AI-generated hallucinations, where AI produces seemingly credible but false information due to its statistical nature and training limitations [7]. - DeepSeek's recent model update reportedly reduced hallucination rates by 45%-50% compared to the previous version [7]. - Despite improvements, the hallucination rate of DeepSeek-R1 was around 21%, ranking fifth among domestic models in a recent evaluation [8]. Group 3: Implications for Media and Information Integrity - The incident serves as a reflection of the fragility of the information ecosystem, emphasizing the need for media organizations to maintain rigor and responsibility in the AI age [8]. - DeepSeek denied any apology regarding the false information, but the generated content still contained inaccuracies [9].