Workflow
提示词工程
icon
Search documents
一句话,性能暴涨49%,马里兰MIT等力作:Prompt才是大模型终极武器
3 6 Ke· 2025-08-18 09:31
Core Insights - The performance improvement of AI models is attributed equally to model upgrades and the optimization of user prompts, with 51% of the enhancement coming from the model and 49% from prompt optimization [2][28]. Group 1: Research Findings - A collaborative study by institutions such as the University of Maryland, MIT, and Stanford demonstrated that user prompts significantly influence AI performance, specifically in image generation tasks using DALL-E models [2][4]. - The concept of "prompt adaptation" was introduced, highlighting the importance of user input in maximizing the capabilities of AI models [3][12]. - The study involved 1,893 participants who generated images using DALL-E 2 and DALL-E 3, revealing that DALL-E 3 outperformed DALL-E 2 due to both model improvements and user prompt adjustments [4][21]. Group 2: Experimental Design - Participants were tasked with generating images based on specific target images, with their performance measured by the cosine similarity between generated and target images [14][15]. - The experiment aimed to separate the effects of model upgrades and prompt optimization on overall performance, using a replay analysis method to assess contributions from both factors [16][26]. - Results indicated that users of DALL-E 3 produced images with a cosine similarity average higher by 0.0164 compared to DALL-E 2 users, demonstrating the model's superior capabilities [22][25]. Group 3: User Behavior and Prompting Strategies - Users of DALL-E 3 tended to create longer and more descriptive prompts, indicating a shift in strategy as they adapted to the model's enhanced capabilities [25][30]. - The study found that the effectiveness of prompt optimization is contingent upon the model's ability to handle complex instructions, suggesting that user input must evolve alongside technological advancements [30][32]. - The research highlighted that lower-skilled users benefited more from model upgrades, while high-skilled users experienced diminishing returns, emphasizing the need for tailored prompting strategies [31][32].
别再空谈“模型即产品”了,AI 已经把产品经理逼到了悬崖边
AI科技大本营· 2025-08-12 09:25
Core Viewpoint - The article discusses the tension between the grand narrative of AI and the practical challenges faced by product managers in implementing AI solutions, highlighting the gap between theoretical concepts and real-world applications [1][2][9]. Group 1: AI Product Development Challenges - Product managers are overwhelmed by the rapid advancements in AI technologies, such as GPT-5 and Kimi K2, while struggling to deliver a successful AI-native product that meets user expectations [1][2]. - There is a significant divide between those discussing the ultimate forms of AGI and those working with unstable model APIs, seeking product-market fit (PMF) [2][3]. - The current AI wave is likened to a "gold rush," where not everyone will find success, and many may face challenges or be eliminated in the process [3]. Group 2: Upcoming Global Product Manager Conference - The Global Product Manager Conference scheduled for August 15-16 aims to address these challenges by bringing together industry leaders to share insights and experiences [2][4]. - Attendees will hear firsthand accounts from pioneers in the AI field, discussing the pitfalls and lessons learned in transforming AI concepts into viable products [5][6]. - The event will feature a live broadcast for those unable to attend in person, allowing broader participation and engagement with the discussions [2][11]. Group 3: Evolving Role of Product Managers - The skills traditionally relied upon by product managers, such as prototyping and documentation, are becoming less relevant due to the rapid evolution of AI technologies [9]. - Future product managers will need to adopt new roles, acting as strategists, directors, and psychologists to navigate the complexities of AI integration and user needs [9][10]. - The article emphasizes the importance of collaboration and networking in this uncertain "great maritime era" of AI development [12].
仅用提示词工程摘下IMO金牌!清华校友强强联手新发现,学术界不靠砸钱也能比肩大厂
量子位· 2025-08-02 05:23
Core Viewpoint - The collaboration between two Tsinghua University alumni has successfully enhanced the Gemini 2.5 Pro model to achieve a gold medal level in the International Mathematical Olympiad (IMO) through a self-iterative verification process and prompt optimization [1][4][10]. Group 1: Model Performance and Methodology - Gemini 2.5 Pro achieved a 31.55% accuracy rate in solving IMO problems, significantly outperforming other models like O3 and Grok 4 [9]. - The research team utilized a structured six-step self-verification process to improve the model's performance, which includes generating initial solutions, self-improvement, and validating solutions [16][18]. - The model was able to generate complete and mathematically rigorous solutions for 5 out of 6 IMO problems, demonstrating the effectiveness of the structured iterative process [24][23]. Group 2: Importance of Prompt Design - The use of specific prompt designs significantly improved the model's ability to solve complex mathematical problems, highlighting the importance of prompt engineering in AI model performance [12][14]. - The research indicated that detailed prompts could reduce the computational search space and enhance efficiency without granting the model new capabilities [23]. Group 3: Research Team Background - The authors, Huang Yichen and Yang Lin, are both Tsinghua University alumni with extensive academic backgrounds in physics and computer science, contributing to the credibility of the research [26][28][33]. - Yang Lin is currently an associate professor at UCLA, focusing on reinforcement learning and generative AI, while Huang Yichen has a strong background in quantum physics and machine learning [30][35]. Group 4: Future Directions and Insights - The research team plans to enhance the model's capabilities through additional training data and fine-tuning, indicating a commitment to ongoing improvement [42]. - Yang Lin expressed the potential for AI to play a more significant role in mathematical research, especially in addressing long-standing unresolved problems [44].
深度评测:PromptPilot,字节跳动的“提示词工厂”
Tai Mei Ti A P P· 2025-08-01 00:27
Core Insights - The article discusses the evolution of prompt engineering in AI, emphasizing its importance in enhancing the interaction between users and AI models [4][16][65] - It highlights the differences in AI model performance based on the quality of prompts used, suggesting that effective prompt engineering can significantly improve AI outputs [3][16][65] Group 1: Evolution of Prompt Engineering - The evolution of prompts has progressed through three stages: "Magic Spell" era, "Enlightenment and Guidance" era, and "Systematic Engineering" era [10][11][14] - In the "Magic Spell" era, users treated AI like a search engine, leading to inconsistent results [10] - The "Enlightenment and Guidance" era introduced techniques like example learning and thinking chains, improving AI's reasoning and logic capabilities [12][13] - The current "Systematic Engineering" era requires structured prompts that include roles, objectives, constraints, examples, and steps to ensure stable and controllable AI outputs [14][15] Group 2: Importance of Prompt Engineering - Prompt engineering is defined as the science of designing and optimizing prompts to effectively communicate with large language models, directly impacting the quality of AI outputs [16] - High-quality prompts reduce the likelihood of AI generating "hallucinations" and help uncover the AI's potential for complex tasks [17] - The R.O.L.E.S. framework (Role, Objective, Limit & Constraint, Examples, Steps) is introduced as a method for creating effective prompts [17][18][20][22][26][28] Group 3: ByteDance's PromptPilot - ByteDance launched PromptPilot, a platform aimed at optimizing the entire process of AI model application, from concept to deployment and iteration [35] - The platform offers features for prompt generation and optimization, making it accessible for users without prior prompt writing experience [39] - Users can validate and refine prompts through various tuning modes, enhancing the effectiveness of AI-generated outputs [40][41][62] Group 4: Conclusion and Future Implications - The article concludes that mastering prompt engineering is essential for leveraging AI effectively, transforming it into a foundational skill for future interactions with AI [65][66] - While PromptPilot is not perfect, it serves as a valuable tool for users to develop structured thinking and improve their interactions with AI [67][70]
AI 产品经理们的挑战:在「审美」之前,都是技术问题
Founder Park· 2025-07-31 03:01
Core Viewpoint - The article discusses the challenges of creating valuable AI Native products, emphasizing that user experience has evolved from a design-centric issue to a technical one, where both user needs and value delivery are at risk of "loss of control" [3][4]. Group 1: User Experience Challenges - The transition from mobile internet to AI Native products has made it more difficult to deliver a valuable user experience, as it now involves complex technical considerations rather than just aesthetic design [3]. - The current bottleneck in AI Native product experience is fundamentally a technical issue, requiring advancements in both product engineering and model technology to reach a market breakthrough [4]. Group 2: Input and Output Dynamics - AI products are structured around the concept of Input > Output, where the AI acts as a "Magic Box" that needs to manage uncertainty effectively [6]. - The focus should be on enhancing the input side to provide better context and clarity, as many users struggle to articulate their needs clearly [7][8]. Group 3: Proposed Solutions - Two key approaches are highlighted: "Context Engineering" by Andrej Karpathy, which emphasizes optimizing the input context for AI, and "Spec-writing" by Sean Grove, which advocates for structured documentation to clarify user intentions [7][8]. - The article argues that the future of AI products should not rely on users becoming experts in context management but rather on AI developing the capability to autonomously understand and predict user intentions [11][12]. Group 4: The Role of AI - The article posits that AI must evolve to become a proactive partner that can interpret and respond to the chaotic nature of human communication and intent, rather than depending on users to provide clear instructions [11][12]. - The ultimate goal is to achieve a "wide input" system that captures high-resolution data from users' lives, creating a feedback loop between input and output for continuous improvement [11].
OpenAI推出学习模式,AI教师真来了?
Hu Xiu· 2025-07-30 01:45
Core Insights - OpenAI has introduced a significant update to ChatGPT called Study Mode, which is designed to enhance the learning experience for users by guiding them through problem-solving rather than just providing answers [1][2]. Group 1: Features of Study Mode - In Study Mode, ChatGPT acts as a mentor, using Socratic questioning and hints to encourage active learning and deeper understanding [3][4]. - The mode is accessible to free users and has received positive feedback for its interactive prompts and structured responses that reduce cognitive load [4][6]. - The system is tailored to individual users based on their skill levels and previous interactions, providing personalized support [4][12]. Group 2: Educational Approach - The underlying framework of Study Mode is developed in collaboration with educators and experts, focusing on core behaviors that promote deeper learning, such as encouraging participation and managing cognitive load [12]. - Key instructional strategies include checking for understanding, reinforcing concepts, and using varied pacing to maintain engagement [20][24]. - The mode emphasizes collaboration with users to help them discover answers rather than providing direct solutions, fostering a more interactive learning environment [25][26].
刚刚,OpenAI推出学习模式,AI教师真来了,系统提示词已泄露
3 6 Ke· 2025-07-30 01:37
Core Insights - OpenAI has introduced a significant update to ChatGPT called Study Mode, which aims to assist users in problem-solving step by step rather than just providing direct answers [1][2]. Features and Characteristics - **Interactive Prompts**: The Study Mode employs Socratic questioning and hints to encourage active learning, rather than simply delivering answers [2]. - **Scaffolding Responses**: Information is organized into easily digestible sections, highlighting key connections between topics to reduce cognitive load [2]. - **Personalized Support**: The mode tailors courses based on users' skill levels and previous interactions, enhancing the learning experience [2]. - **Knowledge Testing**: It includes quizzes and open-ended questions with personalized feedback to track progress and reinforce knowledge [2]. - **Flexibility**: Users can easily switch to Study Mode during conversations, allowing for adaptable learning objectives [2]. Implementation and Design - OpenAI collaborated with educators and experts to develop a custom system of instructions that promote deeper learning behaviors, such as encouraging participation and managing cognitive load [10]. - The system prompts are designed to help users discover answers through guidance rather than direct solutions [13][15]. User Experience - Users can utilize Study Mode for various educational purposes, including homework assistance and exam preparation [4]. - The mode begins by assessing the user's understanding of the topic before providing tailored instructional support [6].
刚刚,OpenAI推出学习模式,AI教师真来了,系统提示词已泄露
机器之心· 2025-07-30 00:48
Core Viewpoint - ChatGPT has introduced a new feature called Study Mode, which aims to enhance user learning by guiding them through problem-solving rather than simply providing answers [1][2][4]. Summary by Sections Features of Study Mode - The Study Mode includes interactive prompts that encourage active learning through Socratic questioning and hints, rather than direct answers [5]. - Responses are organized into understandable sections, highlighting key connections between topics to reduce cognitive load [5]. - The mode offers personalized support tailored to the user's skill level and previous interactions [5]. - Knowledge assessments, including quizzes and open-ended questions, are provided to track progress and reinforce learning [5]. - Users can easily switch to Study Mode during conversations, allowing for flexible learning objectives [5]. User Experience - Initial feedback on the Study Mode has been overwhelmingly positive, indicating its effectiveness in enhancing the learning experience [6]. - A practical example demonstrated how ChatGPT assesses the user's understanding before tailoring the teaching approach to their knowledge level [9]. Development Insights - OpenAI has collaborated with educators and experts to create a system of prompts that support deeper learning behaviors, such as encouraging active participation and providing actionable feedback [13]. - The underlying principles of the Study Mode are based on extensive research in learning sciences [13]. Prompt Engineering - OpenAI has openly shared the key components of the system prompts used in Study Mode, emphasizing the importance of understanding user goals and building on existing knowledge [16][17][18]. - The approach focuses on guiding users through questions and prompts rather than providing direct answers, fostering a collaborative learning environment [19][22].
Karpathy:我不是要造新词,是「上下文工程」对 Agent 来说太重要了
Founder Park· 2025-07-04 13:10
Core Viewpoint - The concept of "Context Engineering" has gained traction in the AI industry, emphasizing that the effectiveness of AI applications relies more on the quality of context provided than on the prompts used to query the AI [1][3]. Group 1: Definition and Importance of Context Engineering - Context Engineering is defined as the discipline of designing and constructing dynamic systems that provide appropriate information and tools to large language models (LLMs) at the right time and in the right format [19]. - The quality of context provided to an AI agent is crucial for its effectiveness, surpassing the complexity of the code or framework used [24]. - A well-constructed context can significantly enhance the performance of AI agents, as demonstrated by examples where rich context leads to more relevant and useful responses [25]. Group 2: Components of Context Engineering - Context Engineering encompasses various elements, including prompt engineering, current state or dialogue history, long-term memory, and retrieval-augmented generation (RAG) [15][11]. - The distinction between prompts, prompt engineering, and context engineering is clarified, with prompts being the immediate instructions given to the AI, while context engineering involves a broader system that dynamically generates context based on task requirements [15][19]. Group 3: Strategies for Implementing Context Engineering - Four common strategies for implementing Context Engineering are identified: writing context, selecting context, compressing context, and isolating context [26]. - Writing context involves saving information outside the context window to assist the agent in completing tasks, such as maintaining a calendar or email history [28][29]. - Selecting context refers to pulling necessary information into the context window to aid the agent, which can include filtering relevant memories or examples [36][38]. - Compressing context focuses on retaining only the essential tokens needed for task execution, often through summarization techniques [43][44]. - Isolating context involves distributing context across multiple agents or using environments to manage context effectively, enhancing task focus and reducing token consumption [47][50].
登上热搜!Prompt不再是AI重点,新热点是Context Engineering
机器之心· 2025-07-03 08:01
Core Viewpoint - The article emphasizes the importance of "Context Engineering" as a systematic approach to optimize the input provided to Large Language Models (LLMs) for better output generation [3][11]. Summary by Sections Introduction to Context Engineering - The article highlights the recent popularity of "Context Engineering," with notable endorsements from figures like Andrej Karpathy and its trending status on platforms like Hacker News and Zhihu [1][2]. Understanding LLMs - LLMs should not be anthropomorphized; they are intelligent text generators without beliefs or intentions [4]. - LLMs function as general, uncertain functions that generate new text based on provided context [5][6][7]. - They are stateless, requiring all relevant background information with each input to maintain context [8]. Focus of Context Engineering - The focus is on optimizing input rather than altering the model itself, aiming to construct the most effective input text to guide the model's output [9]. Context Engineering vs. Prompt Engineering - Context Engineering is a more systematic approach compared to the previously popular "Prompt Engineering," which relied on finding a perfect command [10][11]. - The goal is to create an automated system that prepares comprehensive input for the model, rather than issuing isolated commands [13][17]. Core Elements of Context Engineering - Context Engineering involves building a "super input" toolbox, utilizing various techniques like Retrieval-Augmented Generation (RAG) and intelligent agents [15][19]. - The primary objective is to deliver the most effective information in the appropriate format at the right time to the model [16]. Practical Methodology - The process of using LLMs is likened to scientific experimentation, requiring systematic testing rather than guesswork [23]. - The methodology consists of two main steps: planning from the end goal backward and constructing from the beginning forward [24][25]. - The final output should be clearly defined, and the necessary input information must be identified to create a "raw material package" for the system [26]. Implementation Steps - The article outlines a rigorous process for building and testing the system, ensuring each component functions correctly before final assembly [30]. - Specific testing phases include verifying data interfaces, search functionality, and the assembly of final inputs [30]. Additional Resources - For more detailed practices, the article references Langchain's latest blog and video, which cover the mainstream methods of Context Engineering [29].