Founder Park
Search documents
RTE 开发者社区 Demo Day、S 创上海科创大会,近期优质 AI 活动都在这里
Founder Park· 2025-09-16 13:22
Group 1 - The article highlights several upcoming AI events, including the "AI Creator Carnival" hosted by Silicon Star, which will take place from September 17 to 21, 2025, featuring technology exchanges, product demos, and workshops [2][4] - The Voice Agent Camp organized by the RTE Developer Community will showcase 17 demo projects related to AI voice services, including AI customer service and AI companionship, on September 22, 2025 [5][6] - The S Innovation Shanghai 2025 event, organized by Slush China, will occur on September 23-24, 2025, featuring six stages with discussions on various fields such as green technology and healthcare [6] Group 2 - The Cloud Summit will feature a dedicated exhibition area for Generation Z innovators, showcasing 50 outstanding cases of AI works, aimed at engaging 60,000 attendees from around the world [7][8] - The article provides links for registration and further details on the events, encouraging participation from AI builders, investors, and industry researchers [5][6][9]
2 亿美元 ARR,AI 语音赛道最会赚钱的公司,ElevenLabs 如何做到快速增长?
Founder Park· 2025-09-16 13:22
Core Insights - ElevenLabs has achieved a valuation of $6.6 billion, with the first $100 million in ARR taking 20 months and the second $100 million only taking 10 months [2] - The company is recognized as the fastest-growing AI startup in Europe, operating in a highly competitive AI voice sector [3] - The CEO emphasizes the importance of combining research and product development to ensure market relevance and user engagement [3][4] Company Growth and Strategy - The initial idea for ElevenLabs stemmed from poor movie dubbing experiences in Poland, leading to the realization of the potential in audio technology [4][5] - The company adopted a dual approach of technical development and market validation, initially reaching out to YouTubers to gauge interest in their product [7][8] - A significant pivot occurred when the focus shifted from dubbing to creating a more emotional and natural text-to-speech model based on user feedback [9][10] Product Development and Market Fit - The company did not find product-market fit (PMF) until they shifted their focus to simpler voice generation needs, which resonated more with users [10] - Key milestones in achieving PMF included a viral blog post and successful early user testing, which significantly increased user interest [10] - The company continues to explore ways to ensure long-term value creation for users, indicating that they have not fully settled on PMF yet [10] Competitive Advantages - ElevenLabs maintains a small team structure to enhance execution speed and adaptability, which is seen as a core advantage over larger competitors [3][19] - The company boasts a top-tier research team and a focused approach to voice AI applications, which differentiates it from larger players like OpenAI [16][18] - The CEO believes that the company's product development and execution capabilities provide a competitive edge, especially in creative voice applications [17][18] Financial Performance - ElevenLabs has recently surpassed $200 million in revenue, achieving this milestone in a rapid timeframe [33] - The company aims to continue its growth trajectory, with aspirations to reach $300 million in revenue within a short period [39][40] - The CEO highlights the importance of maintaining a healthy revenue structure while delivering real value to customers [44] Investment and Funding Strategy - The company faced significant challenges in securing initial funding, with over 30 investors rejecting their seed round [64][66] - Each funding round is strategically linked to product developments or user milestones, rather than being announced for the sake of publicity [70] - The CEO emphasizes the importance of not remaining in a perpetual fundraising state, advocating for clear objectives behind each funding announcement [70]
OpenAI发布GPT-5-Codex:独立编码7小时,能动态调整资源,token消耗更少
Founder Park· 2025-09-16 03:24
Core Insights - OpenAI has released a new model specifically designed for programming tasks, named GPT-5-Codex, which is a specialized version of GPT-5 [3][4] - GPT-5-Codex features a "dual-mode" capability, being both fast and reliable, with improved responsiveness for both small and large tasks [5][6] - The model can execute large-scale refactoring tasks for up to 7 hours continuously, showcasing its efficiency [7] Performance and Features - In SWE-bench validation and code refactoring tasks, GPT-5-Codex outperformed the previous model, GPT-5-high, achieving an accuracy rate of 51.3% compared to 33.9% [9][10] - The model dynamically adjusts resource allocation based on task complexity, reducing token consumption by 93.7% for simpler tasks while doubling the processing time for more complex requests [12][13] - GPT-5-Codex has significantly improved code review capabilities, with incorrect comments dropping from 13.7% to 4.4% and high-impact comments increasing from 39.4% to 52.4% [16][18] Integration and User Experience - The model supports multi-modal interactions, including terminal vibe coding, IDE editing, and GitHub integration, catering to various developer preferences [32] - OpenAI emphasizes the importance of "harnessing" the model, integrating it with infrastructure to enable real-world task execution [29][34] - The user experience is enhanced with a response time of less than 1.5 seconds for code completion, crucial for maintaining developer productivity [30] Competitive Landscape - The release of GPT-5-Codex intensifies the competition in the programming AI space, with various domestic and international players developing similar programming agents [45][46] - Notable competitors include Cursor, Gemini CLI, and Claude Code, which focus on execution capabilities and seamless integration with development environments [51][52] - The market is rapidly evolving, with many companies racing to establish their programming AI solutions, indicating a significant shift in software development practices by 2030 [43][54]
张小珺对话OpenAI姚顺雨:生成新世界的系统
Founder Park· 2025-09-15 05:59
Core Insights - The article discusses the evolution of AI, particularly focusing on the transition to the "second half" of AI development, emphasizing the importance of language and reasoning in creating more generalizable AI systems [4][62]. Group 1: AI Evolution and Language - The concept of AI has evolved from rule-based systems to deep reinforcement learning, and now to language models that can reason and generalize across tasks [41][43]. - Language is highlighted as a fundamental tool for generalization, allowing AI to tackle a variety of tasks by leveraging reasoning capabilities [77][79]. Group 2: Agent Systems - The definition of an "Agent" has expanded to include systems that can interact with their environment and make decisions based on reasoning, rather than just following predefined rules [33][36]. - The development of language agents represents a significant shift, as they can perform tasks in more complex environments, such as coding and internet navigation, which were previously challenging for AI [43][54]. Group 3: Task Design and Reward Mechanisms - The article emphasizes the importance of defining effective tasks and environments for AI training, suggesting that the current bottleneck lies in task design rather than model training [62][64]. - A focus on intrinsic rewards, which are based on outcomes rather than processes, is proposed as a key factor for successful reinforcement learning applications [88][66]. Group 4: Future Directions - The future of AI development is seen as a combination of enhancing agent capabilities through better memory systems and intrinsic rewards, as well as exploring multi-agent systems [88][89]. - The potential for AI to generalize across various tasks is highlighted, with coding and mathematical tasks serving as prime examples of areas where AI can excel [80][82].
RAG 的概念很糟糕,让大家忽略了应用构建中最关键的问题
Founder Park· 2025-09-14 04:43
Core Viewpoint - The article emphasizes the importance of Context Engineering in AI development, criticizing the current trend of RAG (Retrieval-Augmented Generation) as a misleading concept that oversimplifies complex processes [5][6][7]. Group 1: Context Engineering - Context Engineering is considered crucial for AI startups, as it focuses on effectively managing the information within the context window during model generation [4][9]. - The concept of Context Rot, where the model's performance deteriorates with an increasing number of tokens, highlights the need for better context management [8][12]. - Effective Context Engineering involves two loops: an internal loop for selecting relevant content for the current context and an external loop for learning to improve information selection over time [7][9]. Group 2: Critique of RAG - RAG is described as a confusing amalgamation of retrieval, generation, and combination, which leads to misunderstandings in the AI community [5][6]. - The article argues that RAG has been misrepresented in the market as merely using embeddings for vector searches, which is seen as a shallow interpretation [5][7]. - The author expresses a strong aversion to the term RAG, suggesting that it detracts from more meaningful discussions about AI development [6][7]. Group 3: Future Directions in AI - Two promising directions for future AI systems are continuous retrieval and remaining within the embedding space, which could enhance performance and efficiency [47][48]. - The potential for models to learn to retrieve information dynamically during generation is highlighted as an exciting area of research [41][42]. - The article suggests that the evolution of retrieval systems may lead to a more integrated approach, where models can generate and retrieve information simultaneously [41][48]. Group 4: Chroma's Role - Chroma is positioned as a leading open-source vector database aimed at facilitating the development of AI applications by providing a robust search infrastructure [70][72]. - The company emphasizes the importance of developer experience, aiming for a seamless integration process that allows users to quickly deploy and utilize the database [78][82]. - Chroma's architecture is designed to be modern and efficient, utilizing distributed systems and a serverless model to optimize performance and cost [75][86].
下周二:Agent 搭建好了,来学学怎么极限控制成本
Founder Park· 2025-09-14 04:43
Core Insights - The integration of AI Agents has become a standard feature in AI products, but the hidden costs associated with their operation, such as multi-turn tool calls and extensive context memory, can lead to significant token consumption [2] Cost Control Strategies - Utilizing fully managed serverless platforms like Cloud Run is an effective way to control costs for AI Agent applications, as it can automatically scale based on request volume and achieve zero cost during idle periods [3][7] - Cloud Run can expand instances from zero to hundreds or thousands within seconds based on real-time request volume, allowing for dynamic scaling that balances stability and cost control [7][9] Upcoming Event - An event featuring Liu Fan, a Google Cloud application modernization expert, will discuss techniques for developing with Cloud Run and achieving extreme cost control [4][9] - The session will include real-world examples demonstrating the powerful scaling capabilities of Cloud Run through monitoring charts that illustrate changes in request volume, instance count, and response latency [9]
数据、IP、境外实体,到底先抓谁?一文讲清 AI 出海合规全流程
Founder Park· 2025-09-12 10:06
Core Viewpoint - The article emphasizes the importance of compliance and legal issues for technology companies and AI startups when expanding internationally after achieving product-market fit [2][4]. Group 1: Pre-Departure Considerations - Before going abroad, companies must conduct a legal country investigation to understand the legal system of the target country, including political stability and legal maturity [7]. - The design of the outbound transaction structure is crucial, with options including greenfield investment, mergers and acquisitions, or simply exporting products, each with its own advantages and risks [8]. - Companies must complete domestic approval procedures before launching outbound projects, including feasibility reports to the National Development and Reform Commission and record-filing with business departments [8][9]. Group 2: Local Compliance Operations - After going abroad, companies need to pay close attention to local operational compliance issues, as regulations can vary significantly even within the same country [9]. - Key compliance issues include foreign investment access, tax registration requirements, and local labor laws, which can differ by state or region [9][11]. Group 3: Domestic Compliance as a Hard Threshold - Evaluating the political and legal environment of the host country is essential, especially for sensitive industries like AI, which face stringent regulations [12]. - Companies must determine if their outbound projects require special approvals or authorizations and consider the stability of promised policies [12][13]. Group 4: Intellectual Property and Data Compliance Challenges - Intellectual property issues are critical for AI companies, particularly regarding the use of open-source code and the potential need for compliance with various licensing agreements [23]. - Data compliance is complex, especially concerning GDPR requirements, which include user rights, consent principles, and data security measures [27][28]. Group 5: Cross-Border Data Transfer - The legality of cross-border data transfers is multifaceted, with different requirements based on the relationship between the data sender and receiver [29]. - Establishing an overseas entity can enhance local trust and compliance but comes with increased costs and management challenges [30]. Group 6: Specific Compliance Considerations for AI Products - AI products targeting children must navigate additional complexities regarding sensitive personal information and parental consent [31]. - Companies using AI-generated code must be cautious of copyright issues and potential liabilities arising from bugs or security flaws in the code [32].
Claude 官方发文:如何给 Agent 构建一个好用的工具?
Founder Park· 2025-09-12 10:06
Core Insights - Anthropic has introduced new features in Claude that allow direct creation and editing of various mainstream office documents, expanding AI's application scenarios in practical tasks [2] - The company emphasizes the importance of designing intuitive tools for uncertain, reasoning AI rather than traditional programming methods [4] - A systematic evaluation of tools using real and complex tasks is essential to validate their effectiveness [5] Group 1 - The focus is on creating integrated workflow tools rather than isolated functionalities, which significantly reduces the reasoning burden on AI [6] - Clear and precise descriptions of tools are crucial for AI to understand their purposes, enhancing the success rate of tool utilization [7] - The article outlines key principles for writing high-quality tools, emphasizing the need for systematic evaluation and collaboration with AI to improve tool performance [13][36] Group 2 - Tools should be designed to reflect the unique affordances of AI agents, allowing them to perceive potential actions differently than traditional software [15][37] - The article suggests building a limited number of well-designed tools targeting high-impact workflows, rather than numerous overlapping functionalities [38] - Naming conventions and namespaces are important for helping AI agents choose the correct tools among many options [40] Group 3 - Tools should return meaningful context to AI, prioritizing high-information signals over technical identifiers to improve task performance [43] - Optimizing tool responses for token efficiency is crucial, with recommendations for pagination and filtering to manage context effectively [48] - The article advocates for prompt engineering in tool descriptions to guide AI behavior and improve performance [52] Group 4 - The future of tool development for AI agents involves shifting from predictable, deterministic patterns to non-deterministic approaches [54] - A systematic, evaluation-driven method is essential for ensuring that tools evolve alongside increasingly powerful AI agents [54]
一文拆解英伟达Rubin CPX:首颗专用AI推理芯片到底强在哪?
Founder Park· 2025-09-12 05:07
Core Viewpoint - Nvidia has launched the Rubin CPX, a CUDA GPU designed for processing large-scale context AI, capable of handling millions of tokens efficiently and quickly [5][4]. Group 1: Product Overview - Rubin CPX is the first CUDA GPU specifically built for processing millions of tokens, featuring 30 petaflops (NVFP4) computing power and 128 GB GDDR7 memory [5][6]. - The GPU can complete million-token level inference in just 1 second, significantly enhancing performance for AI applications [5][4]. - The architecture allows for a division of labor between GPUs, optimizing cost and performance by using GDDR7 instead of HBM [9][12]. Group 2: Performance and Cost Efficiency - The Rubin CPX offers a cost-effective solution, with a single chip costing only 1/4 of the R200 while delivering 80% of its computing power [12][13]. - The total cost of ownership (TCO) in scenarios with long prompts and large batches can drop from $0.6 to $0.06 per hour, representing a tenfold reduction [13]. - Companies investing in Rubin CPX can expect a 50x return on investment, significantly higher than the 10x return from previous models [14]. Group 3: Competitive Landscape - Nvidia's strategy of splitting a general-purpose chip into specialized chips positions it favorably against competitors like AMD, Google, and AWS [15][20]. - The architecture of the Rubin CPX allows for a significant increase in performance, with the potential to outperform existing flagship systems by up to 6.5 times [14][20]. Group 4: Industry Implications - The introduction of Rubin CPX is expected to benefit the PCB industry, as new designs and materials will be required to support the GPU's architecture [24][29]. - The demand for optical modules is anticipated to rise significantly due to the increased bandwidth requirements of the new architecture [30][38]. - The overall power consumption of systems using Rubin CPX is projected to increase, leading to advancements in power supply and cooling solutions [39][40].
算一笔「看不见」的 Agent 成本帐
Founder Park· 2025-09-11 08:25
Core Insights - The integration of AI Agents has become a standard feature in AI products, but the hidden costs associated with their operation pose significant challenges [2] - Controlling costs is crucial, and fully managed serverless platforms like Cloud Run offer a viable solution by automatically scaling based on request volume and achieving zero costs during idle times [3][7] Summary by Sections - **AI Agent Development and Costs** - The deployment of AI Agents is just the initial step, with subsequent operational costs potentially consuming thousands to tens of thousands of tokens per interaction due to multi-turn tool calls and complex logic [2] - **Cost Control Solutions** - Cloud Run is highlighted as an effective platform for managing costs associated with AI Agents, allowing for automatic scaling based on real-time request volume and achieving zero costs when there are no requests [3][7] - **Upcoming Event** - An event featuring Liu Fan, a Google Cloud application modernization expert, will discuss techniques for developing with Cloud Run and strategies for extreme cost control [4][9] - **Key Discussion Points of the Event** - How Cloud Run can scale instances from zero to hundreds or thousands within seconds based on real-time requests [9] - The "zero cost with no requests" model that can reduce the operational costs of AI Agents to zero [9] - Real-world examples demonstrating Cloud Run's scalability through monitoring charts that illustrate changes in request volume, instance count, and response latency [9]