Founder Park
Search documents
数据、IP、境外实体,到底先抓谁?一文讲清 AI 出海合规全流程
Founder Park· 2025-09-12 10:06
产品出海,找到 PMF 之后,下一步就是解决合规和法律问题。 合规的事情,说起来复杂,做起来,也复杂。 数据、知识产权、实体公司、招聘、税务、交易框架、地缘政治…… 听起来就头大。 我们特别邀请到了两位企业出海方面的资深律师,以及 AI 法律类产品的创业者,聊了聊当下科技公 司、AI 创企「出海」面临的合规风险、典型案例及应对方法。 在进行了一些脱敏处理后,Founder Park 整理了本次沉淀内容,很实在的内容,建议收藏。 嘉宾介绍: 超 13000 人的「AI 产品市集」社群!不错过每一款有价值的 AI 应用。 邀请从业者、开发人员和创业者,飞书扫码加群: 李慧君,北京嘉润律师事务所高级合伙人 李然,北京嘉润律师事务所律师 杨帆,WiseLaw 智法数科首席增长官 进群后,你有机会得到: 01 比如,你要在当地聘请当地员工,是否需要有当地实体?或者外派中国员工出去,有没有要求说聘请一 个中国员工就必须按一比一的配比雇佣当地员工?其实每个国家背后的理念是相似的:不仅希望你有个 名头去投资做生意,更希望你的投资能实实在在地造福于他的就业市场或消费者群体,带来新的就业机 会。 产品出海前, 必须要考虑的「四部 ...
Claude 官方发文:如何给 Agent 构建一个好用的工具?
Founder Park· 2025-09-12 10:06
Core Insights - Anthropic has introduced new features in Claude that allow direct creation and editing of various mainstream office documents, expanding AI's application scenarios in practical tasks [2] - The company emphasizes the importance of designing intuitive tools for uncertain, reasoning AI rather than traditional programming methods [4] - A systematic evaluation of tools using real and complex tasks is essential to validate their effectiveness [5] Group 1 - The focus is on creating integrated workflow tools rather than isolated functionalities, which significantly reduces the reasoning burden on AI [6] - Clear and precise descriptions of tools are crucial for AI to understand their purposes, enhancing the success rate of tool utilization [7] - The article outlines key principles for writing high-quality tools, emphasizing the need for systematic evaluation and collaboration with AI to improve tool performance [13][36] Group 2 - Tools should be designed to reflect the unique affordances of AI agents, allowing them to perceive potential actions differently than traditional software [15][37] - The article suggests building a limited number of well-designed tools targeting high-impact workflows, rather than numerous overlapping functionalities [38] - Naming conventions and namespaces are important for helping AI agents choose the correct tools among many options [40] Group 3 - Tools should return meaningful context to AI, prioritizing high-information signals over technical identifiers to improve task performance [43] - Optimizing tool responses for token efficiency is crucial, with recommendations for pagination and filtering to manage context effectively [48] - The article advocates for prompt engineering in tool descriptions to guide AI behavior and improve performance [52] Group 4 - The future of tool development for AI agents involves shifting from predictable, deterministic patterns to non-deterministic approaches [54] - A systematic, evaluation-driven method is essential for ensuring that tools evolve alongside increasingly powerful AI agents [54]
一文拆解英伟达Rubin CPX:首颗专用AI推理芯片到底强在哪?
Founder Park· 2025-09-12 05:07
Core Viewpoint - Nvidia has launched the Rubin CPX, a CUDA GPU designed for processing large-scale context AI, capable of handling millions of tokens efficiently and quickly [5][4]. Group 1: Product Overview - Rubin CPX is the first CUDA GPU specifically built for processing millions of tokens, featuring 30 petaflops (NVFP4) computing power and 128 GB GDDR7 memory [5][6]. - The GPU can complete million-token level inference in just 1 second, significantly enhancing performance for AI applications [5][4]. - The architecture allows for a division of labor between GPUs, optimizing cost and performance by using GDDR7 instead of HBM [9][12]. Group 2: Performance and Cost Efficiency - The Rubin CPX offers a cost-effective solution, with a single chip costing only 1/4 of the R200 while delivering 80% of its computing power [12][13]. - The total cost of ownership (TCO) in scenarios with long prompts and large batches can drop from $0.6 to $0.06 per hour, representing a tenfold reduction [13]. - Companies investing in Rubin CPX can expect a 50x return on investment, significantly higher than the 10x return from previous models [14]. Group 3: Competitive Landscape - Nvidia's strategy of splitting a general-purpose chip into specialized chips positions it favorably against competitors like AMD, Google, and AWS [15][20]. - The architecture of the Rubin CPX allows for a significant increase in performance, with the potential to outperform existing flagship systems by up to 6.5 times [14][20]. Group 4: Industry Implications - The introduction of Rubin CPX is expected to benefit the PCB industry, as new designs and materials will be required to support the GPU's architecture [24][29]. - The demand for optical modules is anticipated to rise significantly due to the increased bandwidth requirements of the new architecture [30][38]. - The overall power consumption of systems using Rubin CPX is projected to increase, leading to advancements in power supply and cooling solutions [39][40].
算一笔「看不见」的 Agent 成本帐
Founder Park· 2025-09-11 08:25
Core Insights - The integration of AI Agents has become a standard feature in AI products, but the hidden costs associated with their operation pose significant challenges [2] - Controlling costs is crucial, and fully managed serverless platforms like Cloud Run offer a viable solution by automatically scaling based on request volume and achieving zero costs during idle times [3][7] Summary by Sections - **AI Agent Development and Costs** - The deployment of AI Agents is just the initial step, with subsequent operational costs potentially consuming thousands to tens of thousands of tokens per interaction due to multi-turn tool calls and complex logic [2] - **Cost Control Solutions** - Cloud Run is highlighted as an effective platform for managing costs associated with AI Agents, allowing for automatic scaling based on real-time request volume and achieving zero costs when there are no requests [3][7] - **Upcoming Event** - An event featuring Liu Fan, a Google Cloud application modernization expert, will discuss techniques for developing with Cloud Run and strategies for extreme cost control [4][9] - **Key Discussion Points of the Event** - How Cloud Run can scale instances from zero to hundreds or thousands within seconds based on real-time requests [9] - The "zero cost with no requests" model that can reduce the operational costs of AI Agents to zero [9] - Real-world examples demonstrating Cloud Run's scalability through monitoring charts that illustrate changes in request volume, instance count, and response latency [9]
Mira Murati 创业公司首发长文,尝试解决 LLM 推理的不确定性难题
Founder Park· 2025-09-11 07:17
Core Insights - The article discusses the challenges of achieving reproducibility in large language model (LLM) inference, highlighting that even with the same input, different outputs can occur due to the probabilistic nature of the sampling process [10][11] - It introduces the concept of "batch invariance" in LLM inference, emphasizing the need for consistent results regardless of batch size or concurrent requests [35][40] Group 1 - Thinking Machines Lab, founded by former OpenAI CTO Mira Murati, has launched a blog series called "Connectionism" to share insights on AI research [3][8] - The blog's first article addresses the non-determinism in LLM inference, explaining that even with a temperature setting of 0, results can still vary [10][12] - The article identifies floating-point non-associativity and concurrency as key factors contributing to the uncertainty in LLM outputs [13][24] Group 2 - The article explains that the assumption of "concurrency + floating-point" as the sole reason for non-determinism is incomplete, as many operations in LLMs can be deterministic [14][16] - It discusses the importance of understanding the implementation of kernel functions in GPUs, which can lead to unpredictable results due to the lack of synchronization among processing cores [25][29] - The article emphasizes that most LLM operations do not require atomic addition, which is often a source of non-determinism, thus allowing for consistent outputs during forward propagation [32][33] Group 3 - The concept of batch invariance is explored, indicating that the results of LLM inference can be affected by the batch size and the order of operations, leading to inconsistencies [36][40] - The article outlines strategies to achieve batch invariance in key operations like RMSNorm, matrix multiplication, and attention mechanisms, ensuring that outputs remain consistent regardless of batch size [42][60][64] - It concludes with a demonstration of deterministic inference using batch-invariant kernel functions, showing that consistent outputs can be achieved with the right implementation [74][78]
Seedream 4.0 来了,AI 图片创业的新机会也来了
Founder Park· 2025-09-11 04:08
Core Viewpoint - The article discusses the emergence of AI image generation models, particularly focusing on the capabilities and advancements of the Seedream 4.0 model developed by Huoshan Engine, which is positioned as a competitive alternative to existing models like Nano Banana and GPT-4o Image [2][4][69]. Group 1: AI Image Generation Models - The AI image generation field has seen significant breakthroughs this year, with models like GPT-4o generating popular images in the Ghibli style [3]. - The Nano Banana model gained attention for its ability to generate high-fidelity images and solve issues related to subject consistency, being compared to ChatGPT in the image generation space [4]. - Huoshan Engine's Seedream 4.0 model offers enhanced capabilities, including multi-image fusion, reference image generation, and image editing, with a focus on improving subject consistency [5][6]. Group 2: Features of Seedream 4.0 - Seedream 4.0 is the first model to support 4K multi-modal image generation, significantly broadening its usability [6]. - The model allows users to input multiple images and generate a high number of outputs simultaneously, showcasing its advanced multi-image fusion capabilities [10][14]. - It supports both single and multi-image inputs, enabling complex creative tasks and maintaining consistency across generated images [50][62]. Group 3: Editing and Customization Capabilities - Seedream 4.0 features strong editing capabilities, allowing users to make precise modifications to images by simply describing the desired changes in natural language [23][24]. - The model can understand and execute detailed instructions, such as replacing elements in an image or adjusting specific details like clothing folds and lighting [26][34]. - It maintains high subject consistency across different creative forms, effectively avoiding common issues like appearance distortion and semantic misalignment during multi-round edits [28][50]. Group 4: Performance and Speed - The model achieves fast image generation speeds, producing images in seconds, which enhances the creative workflow's responsiveness [36]. - With 4K output resolution, Seedream 4.0 delivers high-quality images suitable for commercial publishing, improving detail, color depth, and semantic consistency [39][41]. Group 5: Implications for AI Entrepreneurship - The introduction of context-aware dialogue capabilities in Seedream 4.0 allows for iterative image editing, making it easier for developers to create complex image products without extensive workflow management [69][76]. - This shift in API design enables a more fluid interaction with image generation tools, potentially transforming the landscape of AI image product development [69][70]. - The model's capabilities suggest new entrepreneurial opportunities in the AI image generation space, particularly for products that require iterative design and modification [67][72].
Granola 为什么能赢:会议笔记,把产品做简单很重要
Founder Park· 2025-09-10 12:16
Core Insights - Granola differentiates itself in the crowded "meeting note tool" market by focusing on minimalistic design and leveraging user context effectively [2][3][4] - The primary competitor for Granola is not other AI note-taking products but Apple Notes, as users have only a brief window of 500 milliseconds to decide to take notes during meetings [2][10] Product Design Philosophy - The design philosophy of Granola is based on "lizard brain design," which emphasizes simplicity and minimal intrusion to maximize utility in high-pressure meeting environments [4][9] - Granola aims to be "invisible" during meetings, avoiding the use of intrusive bots that could disrupt the user experience [10][11] User Context and Feedback - Understanding user context is crucial for AI to be helpful, and Granola prioritizes gathering extensive user feedback to inform design decisions [4][14] - The company conducts regular user interviews to ensure they remain aligned with user needs and avoid assumptions about what users want [15] AI Model Utilization - Granola employs the best available third-party AI models initially, only developing proprietary models when necessary to enhance user experience [17][19] - The integration of multiple AI models allows Granola to tailor responses based on user needs and meeting contexts [18][19] Target User Base - Initially, Granola targeted venture capitalists due to their frequent meetings and specific note-taking needs, later expanding to serve founders and other knowledge workers [29][30] - The company believes that if it can effectively serve founders, it can meet the needs of a broader user base [29] Growth Mechanisms - Granola's growth has been driven by user recommendations rather than aggressive marketing strategies, with users often promoting the product in meetings [30][31] - The ability to share notes via links has become a significant growth driver, allowing users to introduce Granola to others seamlessly [30] Future Directions - Granola plans to develop features that allow for cross-meeting analysis and deeper insights based on accumulated context from past meetings [33][36] - The company envisions a future where AI tools can provide real-time insights and recommendations based on a user's entire meeting history [33][36] Competitive Landscape - Granola operates in a competitive landscape where many established players have entered the AI note-taking space, yet it maintains a unique position by focusing on user-centric design [35][38] - The company believes that its approach to creating a personalized tool will allow it to compete effectively against larger firms like OpenAI and Google [38][39]
从 AI 3D生成转型AI原生影视公司,Utopai Studios想「稍微」改造下好莱坞
Founder Park· 2025-09-10 12:16
Core Viewpoint - The article discusses the transformation of Utopai Studios, a company focused on AI-generated content, and its potential impact on the film industry, emphasizing a shift from traditional production paradigms to a more automated and imaginative approach [2][3]. Group 1: Company Transformation - Utopai Studios, previously known for AI 3D generation, has pivoted to content production, launching two AI films [2]. - The studio is led by Cecilia Shen, a young female director, who emphasizes the importance of storytelling and authenticity in their projects [3][5]. Group 2: Film Projects - Utopai's first film, "Cortés," is based on a well-respected story that has been sought after by major studios for decades, showcasing its potential for success [5][7]. - The second project, "Project Space," is an eight-episode sci-fi series that has already pre-sold in the European market [7]. Group 3: AI Integration in Film Production - Utopai aims to create an end-to-end AI production architecture that significantly reduces costs and accelerates the filmmaking process without compromising quality [8]. - The company believes that automation will revolutionize the industry, allowing creators to focus on artistic expression rather than budget constraints [9][15]. Group 4: Challenges and Solutions - Utopai acknowledges the challenges in AI video production, particularly in quality, consistency, and controllability, and is focused on addressing these issues through specialized models [12][13]. - The company’s approach involves integrating 3D data into model training to enhance the understanding of physical interactions in scenes, thus improving the realism of generated content [14]. Group 5: Future of Content Creation - Utopai envisions a future where storytelling and artistic vision take precedence over budgetary limitations, allowing for a broader range of creative projects to be realized [18]. - The integration of AI in filmmaking is seen as a means to democratize the industry, enabling independent creators to produce high-quality content at lower costs [18].
没有法律背景、聊了100位律师后开始创业,他搞出了一家7亿美元估值的AI公司
Founder Park· 2025-09-09 12:53
在垂直 AI 领域,法律科技一直是最受瞩目的赛道之一。 Legora 可以说是这个赛道里增长最快的创企。 和 Lovable 一样,Legora 起源于瑞典,在成立不到两年的时间里,先是拿下了欧洲市场,再向美国市场 扩张,与全球 250 家律所达成了合作,其中不乏Cleary Gottlieb、Goodwin 等顶级律所。 近期,Legora 获得了由 ICONIQ 和 General Catalyst 领投的 8000 万美元 B 轮融资,估值达到 6.75 亿美 元。成为 Harvey 的强劲竞争对手。 而 Legora 的创始人 Max Junestrand 年仅 25 岁,且没有任何法律背景。怎么做成的?创始人的认知是很 关键的。 Max Junestrand 表示,「保持极度的谦逊,谦逊地承认我们不了解这个行业。然后借此机会,与早期合 作伙伴建立关系,每天都进行反馈交流。在一个正在经历巨变的行业中,带着一种更天真的视角反而是 有益的,你会思考,为什么事情要这么做?」 Max Junestrand 最近接受了 YC 合伙人 Gustaf Alströmer 的专访,深入分享了 Legora 背后的思考 ...
Agent 搭起来了,成本怎么控制?
Founder Park· 2025-09-09 12:53
Core Insights - The article discusses the recent trends in AI entrepreneurship, focusing on key areas such as agent development, product internationalization, overseas marketing, and cost control [2][3]. Group 1: Overseas Growth and Profitability - Companies must prioritize profitability from day one when expanding overseas, with a focus on revenue generation [5]. - AI companies exhibit three distinct characteristics: explosive product-driven growth, a mindset of "profit from day zero," and a proactive approach to using AI tools for marketing [5]. Group 2: AI Advertising Innovations - New advertising formats driven by AI are emerging, such as AI Overview and AI Mode, which enhance user experience by integrating ads into AI-generated content [9]. - Traditional marketing strategies are evolving from keyword-based approaches to understanding user intent, leading to more effective ad placements [9]. - AI significantly reduces the costs associated with creative material production, enabling scalable high-quality advertising content [9]. Group 3: AI Agent Development Techniques - The shift in AI agent development from deterministic programming to probabilistic orchestration is crucial, emphasizing the need for agents to understand "what to do" rather than "how to do it" [10]. - Key challenges in developing reliable AI agents include predictability, stability, and operational management (AgentOps) [10][11]. - Effective agent collaboration requires precise definitions of capabilities and skills through well-crafted Agent Cards [12]. Group 4: Cost Control in AI Agent Operations - Upcoming discussions will focus on using Cloud Run for cost-effective AI agent operations, including scaling instances based on real-time demand and achieving zero-cost operation models [15][20]. - Strategies for ensuring agent stability include adopting standardized protocols, implementing retry mechanisms, and maintaining human oversight [16]. - Effective monitoring and management of agent behavior require detailed logging, automated evaluation systems, and the use of tracking tools [16].