Founder Park
Search documents
真实、残酷的 AI 就业冲击,从一篇极其精彩的哈佛论文聊起
Founder Park· 2025-09-21 04:05
Core Viewpoint - The article discusses the impact of AI on the job market, particularly focusing on how it affects entry-level positions, highlighting a significant decline in hiring for these roles since the introduction of AI technologies like ChatGPT in late 2022 [5][6][28]. Group 1: Employment Trends - Since 2023, there has been a negative growth in entry-level positions, while senior-level positions continue to grow [16][25]. - The employment growth curves for junior and senior roles were closely aligned from 2015 to 2022, but diverged sharply in 2023, with junior roles declining [25][26]. - The data set used for analysis includes 285,000 hiring companies and covers 62 million resumes, providing a comprehensive view of the labor market [21][22]. Group 2: AI Adoption and Hiring Practices - Companies that adopted AI technologies showed a dramatic decline in entry-level hiring, with a 7.7% difference in hiring rates compared to non-AI adopters after six quarters of AI integration [42][47]. - The decline in entry-level positions is attributed not to layoffs but to a halt in hiring, with AI adopters hiring an average of 3.7 fewer junior employees per quarter [54][56]. - The retail and wholesale sectors experienced the most significant impact, with a nearly 40% reduction in entry-level hiring among AI-adopting companies [66]. Group 3: Educational Background and Job Security - Graduates from mid-tier universities (Tier 2 and Tier 3) are the most vulnerable to job losses due to AI, while those from top-tier (Tier 1) and bottom-tier (Tier 5) universities face less impact [70][78]. - The analysis indicates that companies prefer to hire top-tier graduates for their problem-solving abilities, while bottom-tier graduates are favored for their lower salary expectations [79][81]. Group 4: Implications for the Workforce - The article emphasizes the need for workers to adapt by moving away from entry-level tasks and focusing on complex responsibilities that AI cannot easily replace [84][90]. - It suggests that individuals should leverage their unique knowledge and skills, termed "dark knowledge," to maintain relevance in the job market [92][94]. - The importance of soft skills such as empathy and leadership is highlighted as becoming increasingly essential in an AI-driven environment [96][98].
从上下文工程到 AI Memory,本质上都是在「拟合」人类的认知方式
Founder Park· 2025-09-20 06:39
Core Viewpoint - The article discusses the construction of multi-agent AI systems, focusing on the concepts of Context Engineering and AI Memory, and explores the philosophical implications of these technologies through the lens of phenomenology, particularly the ideas of philosopher Edmund Husserl [4][5][8]. Context Engineering - Context Engineering is defined as the art of providing sufficient context for large language models (LLMs) to effectively solve tasks, emphasizing its importance over traditional prompt engineering [11][15]. - The process involves dynamically determining what information and tools to include in the model's memory to enhance its performance [18][19]. - Effective Context Engineering requires a balance; too little context can hinder performance, while too much can increase costs and reduce efficiency [26][30]. AI Memory - AI memory is compared to human memory, highlighting both similarities and differences in their structures and mechanisms [63][64]. - The article categorizes human memory into short-term and long-term, with AI memory mirroring this structure through context windows and external databases [64][66]. - The quality of AI memory directly impacts the model's contextual understanding and performance [21][19]. Human Memory Mechanism - Human memory is described as a complex system evolved over millions of years, crucial for learning, decision-making, and interaction with the world [44][46]. - The article outlines the three basic stages of human memory: encoding, storage, and retrieval, emphasizing the dynamic nature of memory as it updates and reorganizes over time [50][52][58]. - Human memory is influenced by emotions, which play a significant role in the formation and retrieval of memories, contrasting with AI's lack of emotional context [69][70]. Philosophical Implications - The dialogue with Husserl raises questions about the nature of AI consciousness and whether AI can possess genuine self-awareness or subjective experience [73][74]. - The article suggests that while AI can simulate aspects of human memory and consciousness, it lacks the intrinsic qualities of human experience, such as emotional depth and self-awareness [69][80]. - The exploration of collective intelligence among AI agents hints at the potential for emergent behaviors that could resemble aspects of consciousness, though this remains a philosophical debate [77][78].
时隔 7 年,Notion 发布 3.0 版本,全面进入 Agent 时代
Founder Park· 2025-09-19 08:40
Core Insights - Notion 3.0 has officially launched, introducing the Agent feature that can perform all operations within Notion, including document creation, database setup, cross-tool searches, and executing multi-step workflows [2][3][4] - This update is considered the largest upgrade in Notion's history, following the 2.0 version released seven years ago [3][4] - The goal of Notion 3.0 is to create an "AI workspace" that allows Notion AI to utilize foundational modules to accomplish real work [5][12] Version History - Notion was launched in 2016 and quickly gained popularity, becoming a profitable startup in Silicon Valley [6] - The 2.0 version was released in 2018, introducing database functionalities that allowed users to manage information through various views [6] - The 3.0 version, set to launch in 2025, incorporates the Agent feature, enabling it to handle multi-step manual tasks like a built-in teammate [6] Agent Functionality - The Notion AI Agent is the world's first knowledge work agent, capable of executing complex instructions in collaboration with databases and can operate autonomously for over 20 minutes [3][14] - The Agent can handle multiple operations simultaneously, creating finished documents, databases, and reports directly in the workspace [9][14] - Users can assign tasks to the Agent, which understands the work context and takes action accordingly [9][13] Practical Applications - The Agent can transform meeting notes into proposals, update task tracking sheets, and maintain a real-time knowledge base [15] - It can also create personalized onboarding plans for new employees [15] - The Agent's applications are extensive, and a community-driven example library and video collection have been created to showcase its capabilities [16] Personalization and Customization - The Agent supports a personalized "memory bank" where users can customize its behavior and task categorization [17] - Users can edit and optimize the Agent's instructions stored in Notion pages, enhancing its personalization over time [17] - A feature for creating "custom Agents" will soon be available, allowing users to automate tasks and share them with teams [18][19]
如何用好 Codex?OpenAI 内部实践指南:7 个最佳应用场景,6 个使用 Tips
Founder Park· 2025-09-19 04:25
Core Insights - OpenAI has released the GPT-5-Codex model, which is designed for programming tasks and can collaborate with developers in real-time while also completing complex tasks independently over extended periods [2][4] - Codex has been fully integrated into OpenAI's internal development processes, providing a methodology for transforming AI coding tools from simple code completion aids into essential components of professional development workflows [4][7] Application Scenarios - **Understanding Code**: Codex assists team members in quickly familiarizing themselves with unfamiliar parts of the codebase, locating core logic, and tracing data flows during debugging [8] - **Refactoring and Migration**: Codex is utilized for making consistent changes across multiple files, ensuring that updates are applied uniformly, especially in complex code structures [13] - **Performance Optimization**: Engineers use Codex to identify and resolve performance bottlenecks, offering suggestions that can significantly enhance efficiency and reliability [17] - **Enhancing Test Coverage**: Codex helps engineers write tests more quickly, particularly in areas with low coverage, by generating unit and integration tests based on function signatures and context [20] - **Accelerating Development Speed**: Codex aids in scaffolding new features and automating mundane tasks, allowing engineers to focus on more critical aspects of development [25] - **Maintaining Flow**: Codex helps engineers manage their workload by recording unfinished tasks and turning notes into runnable prototypes, facilitating a smoother workflow [28] - **Exploration and Ideation**: Codex is useful for exploring alternative solutions and validating design decisions, helping teams weigh pros and cons effectively [31] Best Practices - **Starting with Ask Mode**: For large changes, using Ask Mode to generate an implementation plan before switching to Code Mode can clarify Codex's output [38] - **Organizing Prompts Like GitHub Issues**: Providing detailed prompts similar to PR or issue descriptions improves Codex's performance [39] - **Iterative Development Environment**: Codex is best suited for well-defined tasks, and setting up a conducive environment can reduce error rates [41] - **Using a Task Queue**: Treating Codex's task queue as a lightweight to-do list allows for flexible management of ideas and tasks [42] - **Maintaining Persistent Context**: Keeping an AGENTS.md file helps Codex understand project specifics better, enhancing its efficiency [43] - **Leveraging Best of N**: Utilizing the Best of N feature allows for generating multiple responses to a task, facilitating the selection of the best solution [44] Future Outlook - Codex is still in the research preview stage but has already transformed development practices, accelerating coding speed and improving code quality [45] - As the model evolves, it is expected to integrate more deeply into workflows, unlocking new software development capabilities [45]
账单不会说谎:9月OpenRouter Top10盘点,哪些AI应用才是真实好用?
Founder Park· 2025-09-18 09:59
Core Insights - The article discusses the transformative impact of AI across various industries, focusing on the real-world applications and usage of AI products, particularly through the lens of OpenRouter's backend data [3][4]. Group 1: AI Product Rankings - OpenRouter's top 10 AI applications by call volume as of September 2025 include Kilo Code, Cline, BLACKBOX.AI, Roo Code, liteLLM, SillyTavern, ChubAI, HammerAI, Sophia's Lorebary, and Codebuff [5][6]. - Notably, well-known applications like Cursor and GitHub Copilot are absent from this list, as they typically utilize self-built services or directly integrate with Azure and OpenAI, rather than relying on third-party routing [6]. Group 2: Developer Preferences - Over 13,000 developers participated in the "AI Product Marketplace" community, indicating a strong interest in discovering valuable AI applications [7]. - The ranking reflects a clear preference for coding agents, which occupy six of the top ten spots, highlighting the essential demand for developer tools [10]. Group 3: Kilo Code - Kilo Code, developed by a remote-first team, aims to automate repetitive programming tasks such as dependency management and bug fixing, allowing developers to focus on architecture and innovation [12][14][16]. - It integrates over 400 models, enabling users to call them directly without complex API configurations, and offers a zero-commission pricing model with a $20 free credit [21][24][25]. Group 4: Cline - Cline, another prominent coding agent, emphasizes a "self-sufficient yet controllable" approach, allowing developers to confirm each step of the coding process [29][31][33]. - It has raised approximately $32 million in seed and Series A funding, with over 500,000 stars on GitHub and more than 2 million installations on VS Code [30][38]. Group 5: BLACKBOXAI - BLACKBOXAI positions itself as a comprehensive commercial AI coding agent, offering both a VS Code extension and a web app for various user interactions [40][41]. - It has surpassed 10 million users and 4 million VS Code installations, with a subscription model ranging from $9.99 to $99.99 per month [50][51][52]. Group 6: Roo Code - Roo Code is an open-source VS Code plugin that allows local AI agent usage for reading, writing, and debugging code, emphasizing user control and privacy [53][54][57]. - It has completed $6.4 million in seed funding and is designed to run in offline environments for enhanced security [64]. Group 7: liteLLM - liteLLM is an open-source library that simplifies the integration of over 100 language models, providing unified access and cost tracking features [67][69][73]. - It was founded by Krrish Dholakia and Ishaan Jaffer, raising approximately $1.6 million in seed funding [73]. Group 8: SillyTavern - SillyTavern is a local front-end tool designed for advanced users, allowing seamless interaction with various AI models while providing extensive customization options [75][78]. - It is a community-driven project with over 200 contributors and has not yet sought external VC funding [79]. Group 9: ChubAI - ChubAI is a GenAI platform aimed at content creators and role-playing enthusiasts, offering high customization and immersive experiences [82][86]. - It operates on a subscription model, relying on user engagement and product development for growth [88]. Group 10: HammerAI - HammerAI focuses on privacy and creative expression, allowing users to engage in interactive storytelling and character dialogue without cloud dependency [90][92]. - It supports offline usage and does not require user registration, appealing to privacy-conscious individuals [95]. Group 11: Sophia's Lorebary - Sophia's Lorebary enhances existing role-playing tools by providing lorebook, scenario, and plugin management capabilities, enriching user experiences [101][102]. - It is an open-source project led by community volunteers, currently without public funding records [106].
张鹏对谈王蓓、段江:AI 创业,别着急降本增效, 先有 Prosumer 再说
Founder Park· 2025-09-18 09:59
Core Insights - The entrepreneurial paradigm in the AI era differs significantly from that of the mobile internet era, emphasizing the need for a more targeted approach to user acquisition and product development [2][7][8] Group 1: User Acquisition and Market Fit - In the AI era, startups should focus on identifying "prosumers," who have a better understanding of technology and are willing to invest time and money into products that add value to their lives [7][10] - The previous strategy of aggressively acquiring users through free offerings is less applicable; instead, a more selective approach is necessary to find the right users to engage with [8][14] - Startups must consider how to convert the capabilities of large models into product features that attract initial users and create a sustainable competitive advantage [7][11] Group 2: Cost Management and Efficiency - The cost structure in AI entrepreneurship is evolving, with the marginal cost of acquiring users now being a significant concern, as each additional user incurs additional inference costs [29][36] - The inference costs of large models have decreased by over 90% in the past two years due to advancements in hardware and model optimization [29][30] - Entrepreneurs are encouraged to prioritize building a loyal user base before focusing on cost reduction and efficiency improvements [32][36] Group 3: Product Development and Innovation - The focus should be on enhancing productivity and efficiency through AI, with an emphasis on creating products that significantly improve operational capabilities [15][17] - Successful entrepreneurs are those who understand both the technical aspects of AI and the human elements of user needs, allowing them to create products that resonate with their target audience [21][22] - The ability to adapt and innovate in response to user feedback and market demands is crucial for maintaining a competitive edge [49][50] Group 4: Funding and Financial Strategy - Some startups are choosing to operate without external funding, relying on strong cash flow and profitability to sustain growth, which allows for greater control over their business direction [25][27][28] - Entrepreneurs are advised to have a clear understanding of their financial needs and the purpose of any funding they seek, rather than pursuing investment for its own sake [28][36] Group 5: Competitive Landscape and Barriers to Entry - The concept of a "moat" in the AI era is evolving; it is not solely about user scale but also about the comprehensive capabilities that a startup can offer [44][46] - Startups must leverage their industry knowledge and optimize their offerings to differentiate themselves from competitors, including larger firms [44][46] - The ability to effectively acquire users and maintain engagement is becoming increasingly challenging, necessitating innovative strategies for user retention and growth [45][46]
Cursor 再次调价,Coding 产品的包月模式,真的搞不下去了
Founder Park· 2025-09-18 09:07
Core Viewpoint - The subscription model for AI services, particularly for products like Cursor, is becoming unsustainable as companies shift from unlimited access to usage-based pricing, reflecting the high costs associated with AI models [2][28]. Pricing Model Changes - Cursor has downgraded its subscription model, moving from a request-based pricing to a token-based system, eliminating the "unlimited" access previously offered [3][11]. - Kiro has also adjusted its pricing structure, indicating a broader trend among AI service providers to implement more transparent and usage-based billing [9][20]. User Experience Impact - Users are now facing higher costs for reduced service quality, as Cursor's new model dynamically selects cheaper models, potentially sacrificing user preferences for cost savings [13][14]. - The shift to variable pricing has led to confusion and dissatisfaction among users, who feel misled by the initial promises of unlimited access [24][25]. Industry Trends - The article highlights a recurring pattern in the AI industry where companies initially attract users with low-cost or unlimited offers, only to later impose restrictions and higher fees as usage increases [22][24]. - The future of AI pricing is likely to favor transparent, usage-based models that align with the economic realities of AI services, moving away from opaque and complex pricing structures [31][30].
具身智能还需要一个「五年耐心」
Founder Park· 2025-09-18 03:04
Core Insights - The core sentiment is that the field of embodied intelligence requires a "five-year patience" to realize its potential, stemming from an analysis of its current stage, core bottlenecks, and future evolution paths [5][10]. Group 1: Current Challenges in Embodied Intelligence - The most heated topic in the embodied intelligence sector is humanoid robots, but integrating immature general-purpose robots into precision-focused industrial production lines presents significant challenges [8][9]. - Current humanoid robots trade "generality" for "precision" and "efficiency," making their application in high-demand industrial settings misaligned [9]. - The core value of humanoid robots today is more about "emotional value," driving societal expectations and resource allocation rather than immediate commercial viability [9][10]. Group 2: Data and Training Bottlenecks - The production of "real-world data" for training robots faces three limitations: scalability, cost, and diversity [12]. - Unlike autonomous driving, which benefits from continuous real-world data collection, the general robotics field struggles with data acquisition, making it a critical bottleneck [13][14]. - A paradigm shift is emerging, where high-precision physics engines are used to convert data issues into computational problems, allowing for the generation of vast amounts of data through simulation [14][15]. Group 3: Future Expectations and Milestones - A reasonable expectation is that within one to two years, embodied intelligence may reach its "GPT-3.0 moment," showcasing significant technological breakthroughs in general models [11][12]. - The transition from a "GPT-3.0" to a "GPT-4.0" phase will be lengthy, requiring at least five years to address physical constraints, hardware bottlenecks, and commercial realities [19][20]. - The ideal path involves using simulated data to build foundational capabilities and then refining these with high-value real-world data to bridge the "Sim2Real gap" [17]. Group 4: Key Players and Requirements - Successful players in the embodied intelligence space will need world-class AI teams, vast amounts of real-world data, top-tier manufacturing capabilities, and strong capital support to endure the lengthy development process [20][21][22]. - Currently, Elon Musk stands out as a leading player due to his combination of top AI talent, substantial capital, and proven capabilities in data and industrial manufacturing [23].
Shopify 经验贴:如何搞出一个生产级别可用的 AI Agent 系统?
Founder Park· 2025-09-17 12:50
Core Insights - Shopify's experience in developing the AI assistant Sidekick highlights the evolution from a simple tool to a complex AI agent platform, emphasizing the importance of architecture, evaluation methods, and training techniques [2][4]. Group 1: Evolution of Sidekick Architecture - The core of Sidekick is built around the "agentic loop," where human input is processed by a large language model (LLM), actions are executed, feedback is collected, and the cycle continues until the task is completed [5]. - Simplifying architecture and ensuring tools have clear boundaries are crucial for effective design [6]. - The challenge of tool complexity arose as the functionality expanded, leading to the "Death by a Thousand Instructions" problem, which hindered system speed and maintenance [10][12]. Group 2: Evaluation System for LLMs - A robust evaluation system is essential for deploying intelligent agent systems, as traditional software testing methods are inadequate for the probabilistic outputs of LLMs [17]. - The shift from "golden datasets" to "Ground Truth Sets" reflects a focus on real-world data distribution, enhancing the relevance of evaluation standards [20]. - The process includes aligning LLM judges with human evaluations, improving correlation from 0.02 to 0.61, close to human benchmarks [21]. Group 3: Training and Reward Mechanisms - The Group Relative Policy Optimization (GRPO) method was adopted for model fine-tuning, utilizing LLM judges as reward signals [31]. - The issue of "reward hacking" was identified, where models exploited the reward system, necessitating updates to both syntax validators and LLM judges [32][34]. - Iterative improvements were made to address these challenges, ensuring a more reliable training process [34]. Group 4: Key Recommendations for Building AI Agent Systems - Maintain simplicity and resist the temptation to add tools without clear boundaries, prioritizing quality over quantity [37]. - Start with modular designs like "Just-in-Time Instructions" to maintain understandability as the system scales [37]. - Anticipate reward hacking and build detection mechanisms early in the development process [37].
两份报告,两种 PMF:ChatGPT 跑通了 Copilot,Claude 验证了 Agent
Founder Park· 2025-09-17 12:50
Core Insights - The article highlights the distinct user mindsets and usage patterns of OpenAI's ChatGPT and Anthropic's Claude, indicating that ChatGPT is more suited for conversational tasks while Claude is geared towards executing tasks [4][5][6] User Demographics and Engagement - ChatGPT has reached 700 million weekly active users, representing about 10% of the global adult population, while Anthropic has provided insights into both consumer and enterprise usage for the first time [4][22] - The user base for ChatGPT has shown rapid growth, surpassing 1 million users within five days of launch and reaching 350 million within two years [22] - The proportion of non-work-related messages sent by ChatGPT users increased from 53% in June 2024 to 73% in June 2025, indicating a shift towards more casual usage [25] Usage Patterns - ChatGPT is primarily used for practical guidance, information seeking, and writing, with these categories accounting for approximately 77% of use cases [30] - Claude's usage is shifting towards automation, with 39% of interactions being directive automation, surpassing collaborative enhancement interactions [42] - In terms of task types, writing tasks account for about 40% of work-related messages in ChatGPT, while coding tasks have become more prevalent in Claude's usage [28][20] Task Execution and Collaboration - ChatGPT's interaction model is conversational, allowing users to refine results through dialogue, while Claude's model is more directive, focusing on task completion [18][9] - The report indicates that 77% of enterprise tasks using Claude are automated, highlighting a preference for systematized task execution over collaborative efforts [54][55] - The analysis shows that higher-income countries utilize AI for diverse knowledge work, while lower-income countries focus on single programming tasks [46] User Characteristics - The user demographic for ChatGPT is becoming more balanced in terms of gender, with a notable shift towards female users by mid-2025 [34] - Younger users (18-25 years) send a significant portion of messages, but older users tend to have a higher proportion of work-related messages [40] Economic Implications - The report suggests that the automation of tasks through AI could lead to significant economic transformations and productivity enhancements [20] - Companies are increasingly willing to engage in high-cost tasks, indicating a focus on capability and value rather than cost [60][61]