Workflow
Tinker Cookbook
icon
Search documents
Thinking Machines 发布 Tinker API,实现灵活的模型微调
AI前线· 2025-10-13 13:54
Core Insights - Thinking Machines has launched Tinker, an API designed for fine-tuning open-weight language models, aimed at reducing infrastructure costs for developers [2][5] - Tinker supports various model architectures, allowing developers to fine-tune models with simple Python code modifications [2][3] - The platform integrates LoRA to enhance GPU memory utilization during parallel fine-tuning, making it practical for research teams with limited resources [2] Summary by Sections Tinker API - Tinker provides managed scheduling, GPU allocation, and checkpoint handling, abstracting cluster management for developers [2] - It offers low-level primitives like forward_backward and sample, enabling developers to create new methods without managing infrastructure [3] Tinker Cookbook - The Tinker Cookbook is an open-source repository that implements common fine-tuning techniques, including reinforcement learning methods and preference optimization workflows [3] - Early users from prestigious institutions have applied Tinker to tasks such as theorem proving and multi-agent reinforcement learning [3] Community Feedback - Initial community feedback highlights a balance between flexibility and simplicity, with professionals noting that RLaaS (Reinforcement Learning as a Service) addresses a significant gap for enterprises [4] Founder Insights - The founder of Thinking Machines emphasizes that Tinker provides cutting-edge tools for researchers, simplifying the complexity of distributed training while supporting innovative research and model customization [5] - Tinker is currently in closed testing, with early access being free and a pay-per-use model planned for the future [5]
开发者狂喜:Thinking Machines发布首款产品Tinker,后训练麻烦全给包了
机器之心· 2025-10-02 03:12
Core Insights - Tinker, the first product launched by Thinking Machines, is an API designed to simplify the fine-tuning of language models for developers and researchers, allowing them to focus on training data and algorithms while Tinker manages infrastructure-related tasks [2][4][16]. Product Features - Tinker supports various advanced models, including Qwen-235B-A22B, and allows users to switch from small to large models with ease, akin to changing a string in Python code [6][8]. - The API provides low-level primitives such as forward_backward and sample, which are essential for most common post-training methods. An open-source library, Tinker Cookbook, is also available to offer modern implementations of post-training methods [9][11]. Use Cases and Adoption - Teams from prestigious institutions like Princeton, Stanford, and UC Berkeley are already utilizing Tinker, demonstrating its versatility in supporting both supervised fine-tuning and experimental reinforcement learning pipelines [13]. - The Goedel team at Princeton achieved comparable performance to full-parameter models using only 20% of the data, while Stanford's chemistry group improved accuracy from 15% to 50% in a specific task using Tinker [14]. Market Position and Future Outlook - Tinker aims to democratize access to fine-tuning capabilities, potentially leading to more diverse product innovations in the AI space [16]. - The initial phase of Tinker will be free, with a usage-based pricing model to be introduced in the coming weeks [15].