提示词工程

Search documents
GPT-5差评启示录:用户与AI交互方式还停留在上一个时代
3 6 Ke· 2025-08-21 08:49
Core Insights - GPT-5 has received mixed reviews since its launch on August 8, with users expressing dissatisfaction despite its technical advancements [1][5][7] - The official stance from OpenAI is that the issues stem from users not adapting to the new interaction model required by GPT-5, which has evolved into a more autonomous "digital mind" [9][78] - The release of a prompt guide by OpenAI aims to help users better engage with GPT-5, emphasizing the importance of updated communication methods [8][9] Group 1: Performance and Capabilities - GPT-5 demonstrates significant improvements in areas such as mathematics, coding, and multi-modal understanding, showcasing its capabilities as a "full-stack engineer" [4][13] - Despite its higher IQ, GPT-5 exhibits instability, sometimes making errors on simple tasks and lacking emotional intelligence, which has led to concerns about its practical usability [5][6][10] - OpenAI has reported a performance increase in the Tau-Bench test, with scores rising from 73.9% to 78.2%, indicating better efficiency and lower costs [23][24] Group 2: User Interaction and Guidelines - The prompt guide outlines four key areas of evolution for GPT-5: agentic task performance, coding ability, raw intelligence, and steerability, which are crucial for effective user interaction [10][15][17] - Users are encouraged to adjust parameters like reasoning effort and verbosity to optimize GPT-5's performance based on task complexity [53][70] - The guide suggests methods for users to either constrain or empower GPT-5's capabilities, depending on the task requirements, highlighting the need for a more nuanced approach to AI interaction [29][32][36] Group 3: Challenges and Solutions - The dual-edged nature of GPT-5's capabilities means that improper use can lead to inefficiencies, necessitating users to become adept "trainers" of the AI [26][27] - OpenAI emphasizes the importance of clear and structured prompts to avoid conflicts that could lead to performance degradation [54][56] - The guide provides practical solutions for common user challenges, such as managing verbosity and reasoning depth, to enhance the overall interaction experience [50][52][68]
“现在读AI博士已经太晚了”
量子位· 2025-08-19 05:25
Core Viewpoint - The article discusses the perspective of Jad Tarifi, a founding member of Google's generative AI team, who advises against pursuing a PhD in AI due to the rapid evolution of the field, suggesting that by the time one graduates, the AI landscape may have drastically changed [1][8]. Group 1: AI Talent Market - Major tech companies like Meta are offering signing bonuses reaching hundreds of millions to attract AI talent [2]. - Tarifi's comments serve as a stark contrast to the ongoing talent war in the AI sector, highlighting the urgency and volatility of the field [3][4]. - The job market is being reshaped by AI, with over 1 million jobs in the U.S. announced for layoffs due to generative AI adoption in 2025 alone [14][15]. Group 2: Employment Impact - The technology sector has been particularly affected, with over 89,000 layoffs attributed directly to AI-driven redundancies since 2023 [16]. - Entry-level positions, especially in knowledge-intensive roles, are at risk as AI can perform tasks traditionally handled by junior employees [19]. - Nearly half of U.S. Gen Z job seekers feel that AI has devalued their degrees, reflecting a significant shift in the job market [21]. Group 3: Future Skills and Adaptation - Tarifi emphasizes the importance of developing social skills and empathy as essential competencies in the AI era [23]. - He suggests that while technical knowledge is valuable, understanding how to effectively use AI tools and having a good sense of taste in their application is crucial [24]. - The article also notes that individuals should focus on excelling in specific areas rather than trying to master every detail of AI technology [28].
别再空谈“模型即产品”了,AI 已经把产品经理逼到了悬崖边
AI科技大本营· 2025-08-12 09:25
Core Viewpoint - The article discusses the tension between the grand narrative of AI and the practical challenges faced by product managers in implementing AI solutions, highlighting the gap between theoretical concepts and real-world applications [1][2][9]. Group 1: AI Product Development Challenges - Product managers are overwhelmed by the rapid advancements in AI technologies, such as GPT-5 and Kimi K2, while struggling to deliver a successful AI-native product that meets user expectations [1][2]. - There is a significant divide between those discussing the ultimate forms of AGI and those working with unstable model APIs, seeking product-market fit (PMF) [2][3]. - The current AI wave is likened to a "gold rush," where not everyone will find success, and many may face challenges or be eliminated in the process [3]. Group 2: Upcoming Global Product Manager Conference - The Global Product Manager Conference scheduled for August 15-16 aims to address these challenges by bringing together industry leaders to share insights and experiences [2][4]. - Attendees will hear firsthand accounts from pioneers in the AI field, discussing the pitfalls and lessons learned in transforming AI concepts into viable products [5][6]. - The event will feature a live broadcast for those unable to attend in person, allowing broader participation and engagement with the discussions [2][11]. Group 3: Evolving Role of Product Managers - The skills traditionally relied upon by product managers, such as prototyping and documentation, are becoming less relevant due to the rapid evolution of AI technologies [9]. - Future product managers will need to adopt new roles, acting as strategists, directors, and psychologists to navigate the complexities of AI integration and user needs [9][10]. - The article emphasizes the importance of collaboration and networking in this uncertain "great maritime era" of AI development [12].
仅用提示词工程摘下IMO金牌!清华校友强强联手新发现,学术界不靠砸钱也能比肩大厂
量子位· 2025-08-02 05:23
Core Viewpoint - The collaboration between two Tsinghua University alumni has successfully enhanced the Gemini 2.5 Pro model to achieve a gold medal level in the International Mathematical Olympiad (IMO) through a self-iterative verification process and prompt optimization [1][4][10]. Group 1: Model Performance and Methodology - Gemini 2.5 Pro achieved a 31.55% accuracy rate in solving IMO problems, significantly outperforming other models like O3 and Grok 4 [9]. - The research team utilized a structured six-step self-verification process to improve the model's performance, which includes generating initial solutions, self-improvement, and validating solutions [16][18]. - The model was able to generate complete and mathematically rigorous solutions for 5 out of 6 IMO problems, demonstrating the effectiveness of the structured iterative process [24][23]. Group 2: Importance of Prompt Design - The use of specific prompt designs significantly improved the model's ability to solve complex mathematical problems, highlighting the importance of prompt engineering in AI model performance [12][14]. - The research indicated that detailed prompts could reduce the computational search space and enhance efficiency without granting the model new capabilities [23]. Group 3: Research Team Background - The authors, Huang Yichen and Yang Lin, are both Tsinghua University alumni with extensive academic backgrounds in physics and computer science, contributing to the credibility of the research [26][28][33]. - Yang Lin is currently an associate professor at UCLA, focusing on reinforcement learning and generative AI, while Huang Yichen has a strong background in quantum physics and machine learning [30][35]. Group 4: Future Directions and Insights - The research team plans to enhance the model's capabilities through additional training data and fine-tuning, indicating a commitment to ongoing improvement [42]. - Yang Lin expressed the potential for AI to play a more significant role in mathematical research, especially in addressing long-standing unresolved problems [44].
AI 产品经理们的挑战:在「审美」之前,都是技术问题
Founder Park· 2025-07-31 03:01
Core Viewpoint - The article discusses the challenges of creating valuable AI Native products, emphasizing that user experience has evolved from a design-centric issue to a technical one, where both user needs and value delivery are at risk of "loss of control" [3][4]. Group 1: User Experience Challenges - The transition from mobile internet to AI Native products has made it more difficult to deliver a valuable user experience, as it now involves complex technical considerations rather than just aesthetic design [3]. - The current bottleneck in AI Native product experience is fundamentally a technical issue, requiring advancements in both product engineering and model technology to reach a market breakthrough [4]. Group 2: Input and Output Dynamics - AI products are structured around the concept of Input > Output, where the AI acts as a "Magic Box" that needs to manage uncertainty effectively [6]. - The focus should be on enhancing the input side to provide better context and clarity, as many users struggle to articulate their needs clearly [7][8]. Group 3: Proposed Solutions - Two key approaches are highlighted: "Context Engineering" by Andrej Karpathy, which emphasizes optimizing the input context for AI, and "Spec-writing" by Sean Grove, which advocates for structured documentation to clarify user intentions [7][8]. - The article argues that the future of AI products should not rely on users becoming experts in context management but rather on AI developing the capability to autonomously understand and predict user intentions [11][12]. Group 4: The Role of AI - The article posits that AI must evolve to become a proactive partner that can interpret and respond to the chaotic nature of human communication and intent, rather than depending on users to provide clear instructions [11][12]. - The ultimate goal is to achieve a "wide input" system that captures high-resolution data from users' lives, creating a feedback loop between input and output for continuous improvement [11].
刚刚,OpenAI推出学习模式,AI教师真来了,系统提示词已泄露
3 6 Ke· 2025-07-30 01:37
Core Insights - OpenAI has introduced a significant update to ChatGPT called Study Mode, which aims to assist users in problem-solving step by step rather than just providing direct answers [1][2]. Features and Characteristics - **Interactive Prompts**: The Study Mode employs Socratic questioning and hints to encourage active learning, rather than simply delivering answers [2]. - **Scaffolding Responses**: Information is organized into easily digestible sections, highlighting key connections between topics to reduce cognitive load [2]. - **Personalized Support**: The mode tailors courses based on users' skill levels and previous interactions, enhancing the learning experience [2]. - **Knowledge Testing**: It includes quizzes and open-ended questions with personalized feedback to track progress and reinforce knowledge [2]. - **Flexibility**: Users can easily switch to Study Mode during conversations, allowing for adaptable learning objectives [2]. Implementation and Design - OpenAI collaborated with educators and experts to develop a custom system of instructions that promote deeper learning behaviors, such as encouraging participation and managing cognitive load [10]. - The system prompts are designed to help users discover answers through guidance rather than direct solutions [13][15]. User Experience - Users can utilize Study Mode for various educational purposes, including homework assistance and exam preparation [4]. - The mode begins by assessing the user's understanding of the topic before providing tailored instructional support [6].
刚刚,OpenAI推出学习模式,AI教师真来了,系统提示词已泄露
机器之心· 2025-07-30 00:48
Core Viewpoint - ChatGPT has introduced a new feature called Study Mode, which aims to enhance user learning by guiding them through problem-solving rather than simply providing answers [1][2][4]. Summary by Sections Features of Study Mode - The Study Mode includes interactive prompts that encourage active learning through Socratic questioning and hints, rather than direct answers [5]. - Responses are organized into understandable sections, highlighting key connections between topics to reduce cognitive load [5]. - The mode offers personalized support tailored to the user's skill level and previous interactions [5]. - Knowledge assessments, including quizzes and open-ended questions, are provided to track progress and reinforce learning [5]. - Users can easily switch to Study Mode during conversations, allowing for flexible learning objectives [5]. User Experience - Initial feedback on the Study Mode has been overwhelmingly positive, indicating its effectiveness in enhancing the learning experience [6]. - A practical example demonstrated how ChatGPT assesses the user's understanding before tailoring the teaching approach to their knowledge level [9]. Development Insights - OpenAI has collaborated with educators and experts to create a system of prompts that support deeper learning behaviors, such as encouraging active participation and providing actionable feedback [13]. - The underlying principles of the Study Mode are based on extensive research in learning sciences [13]. Prompt Engineering - OpenAI has openly shared the key components of the system prompts used in Study Mode, emphasizing the importance of understanding user goals and building on existing knowledge [16][17][18]. - The approach focuses on guiding users through questions and prompts rather than providing direct answers, fostering a collaborative learning environment [19][22].
Karpathy:我不是要造新词,是「上下文工程」对 Agent 来说太重要了
Founder Park· 2025-07-04 13:10
Core Viewpoint - The concept of "Context Engineering" has gained traction in the AI industry, emphasizing that the effectiveness of AI applications relies more on the quality of context provided than on the prompts used to query the AI [1][3]. Group 1: Definition and Importance of Context Engineering - Context Engineering is defined as the discipline of designing and constructing dynamic systems that provide appropriate information and tools to large language models (LLMs) at the right time and in the right format [19]. - The quality of context provided to an AI agent is crucial for its effectiveness, surpassing the complexity of the code or framework used [24]. - A well-constructed context can significantly enhance the performance of AI agents, as demonstrated by examples where rich context leads to more relevant and useful responses [25]. Group 2: Components of Context Engineering - Context Engineering encompasses various elements, including prompt engineering, current state or dialogue history, long-term memory, and retrieval-augmented generation (RAG) [15][11]. - The distinction between prompts, prompt engineering, and context engineering is clarified, with prompts being the immediate instructions given to the AI, while context engineering involves a broader system that dynamically generates context based on task requirements [15][19]. Group 3: Strategies for Implementing Context Engineering - Four common strategies for implementing Context Engineering are identified: writing context, selecting context, compressing context, and isolating context [26]. - Writing context involves saving information outside the context window to assist the agent in completing tasks, such as maintaining a calendar or email history [28][29]. - Selecting context refers to pulling necessary information into the context window to aid the agent, which can include filtering relevant memories or examples [36][38]. - Compressing context focuses on retaining only the essential tokens needed for task execution, often through summarization techniques [43][44]. - Isolating context involves distributing context across multiple agents or using environments to manage context effectively, enhancing task focus and reducing token consumption [47][50].
登上热搜!Prompt不再是AI重点,新热点是Context Engineering
机器之心· 2025-07-03 08:01
Core Viewpoint - The article emphasizes the importance of "Context Engineering" as a systematic approach to optimize the input provided to Large Language Models (LLMs) for better output generation [3][11]. Summary by Sections Introduction to Context Engineering - The article highlights the recent popularity of "Context Engineering," with notable endorsements from figures like Andrej Karpathy and its trending status on platforms like Hacker News and Zhihu [1][2]. Understanding LLMs - LLMs should not be anthropomorphized; they are intelligent text generators without beliefs or intentions [4]. - LLMs function as general, uncertain functions that generate new text based on provided context [5][6][7]. - They are stateless, requiring all relevant background information with each input to maintain context [8]. Focus of Context Engineering - The focus is on optimizing input rather than altering the model itself, aiming to construct the most effective input text to guide the model's output [9]. Context Engineering vs. Prompt Engineering - Context Engineering is a more systematic approach compared to the previously popular "Prompt Engineering," which relied on finding a perfect command [10][11]. - The goal is to create an automated system that prepares comprehensive input for the model, rather than issuing isolated commands [13][17]. Core Elements of Context Engineering - Context Engineering involves building a "super input" toolbox, utilizing various techniques like Retrieval-Augmented Generation (RAG) and intelligent agents [15][19]. - The primary objective is to deliver the most effective information in the appropriate format at the right time to the model [16]. Practical Methodology - The process of using LLMs is likened to scientific experimentation, requiring systematic testing rather than guesswork [23]. - The methodology consists of two main steps: planning from the end goal backward and constructing from the beginning forward [24][25]. - The final output should be clearly defined, and the necessary input information must be identified to create a "raw material package" for the system [26]. Implementation Steps - The article outlines a rigorous process for building and testing the system, ensuring each component functions correctly before final assembly [30]. - Specific testing phases include verifying data interfaces, search functionality, and the assembly of final inputs [30]. Additional Resources - For more detailed practices, the article references Langchain's latest blog and video, which cover the mainstream methods of Context Engineering [29].
论坛| 未可知 x 容诚: AI技术在基金行业的创新应用与效率提升之道
未可知人工智能研究院· 2025-07-02 12:01
Core Viewpoint - The application of AI technology in the fund industry is transforming operations and enhancing efficiency, with a focus on generative AI and its capabilities in content production and task completion [1][4]. Group 1: AI Technology Development - The evolution of AI technology has been systematically reviewed, highlighting the essential differences between generative AI and traditional decision-making AI [4]. - Generative AI, represented by tools like DeepSeek and Sora, is reshaping content production methods and enabling a leap from "answering questions" to "completing tasks" [4]. Group 2: Specific Applications in the Fund Industry - Three main directions for efficiency improvement in the fund industry have been identified: 1. High efficiency in information processing, with tools like Secret Tower AI reducing information collection time by 80% [6]. 2. Automation of content production, utilizing prompt engineering to quickly generate marketing copy and presentations [6]. 3. Intelligent business processes, where digital employees can accurately perform repetitive tasks such as net asset value verification [6]. - A case study from a large fund company demonstrated that deploying RPA digital employees led to the automation of most operational processes, saving over 4,000 work hours annually [6]. Group 3: Current State of AI Development in China - The challenges of computational power bottlenecks in China's AI development were acknowledged, alongside the unique advantages of domestic models [8]. - DeepSeek's open-source strategy and low-cost training characteristics provide a cost-effective AI transformation solution for financial institutions [8]. - Emphasis was placed on the importance of data security, with recommendations for localized deployment to address privacy concerns [8]. Group 4: Future Trends and Initiatives - A series of AI training courses will be launched to help financial institutions cultivate AI talent, emphasizing that the next decade will be a golden period for human-machine collaboration [13]. - Institutions that can build "AI employee" teams early will gain a competitive edge in the industry [13]. - The presentation provided a clear roadmap for the digital transformation of the fund industry, combining theoretical insights with practical value [13].