Fine-tuning
Search documents
The Powerful Alternative To Fine-Tuning
Y Combinator· 2026-02-27 15:00
The world is changing so quickly. This is probably a little bit obvious, but you should just try things and and like every day do something with AI. Last summer, I took a weekend and used um GPT5 to help me build an iPhone app.I hadn't done that in a decade. And yeah, it's so fast and so easy. And that was, you know, an age ago. That was like 8 months ago.Uh now it's even faster and easier. Don't limit yourself. like anything that you imagine, you should just try to use AI and see how far you can get with i ...
Inside the AI Factory: How DDN Powers End-to-End AI Workflows
DDN· 2026-01-10 00:49
[MUSIC] Hello, I'm Jason from DDN. Today, I want to take you inside the AI Factory and show you how DDN powers full cycle AI workflows from the moment data arrives all the way through training, tuning, and inference. When people talk about AI, they often think only about model training. But an AI Factory is a pipeline with many unique stages, with ingest, data preparation, model training, fine-tuning, and then inference and RAG. As data moves through these stages, the volume increases, the access pattern ...
X @Avi Chawla
Avi Chawla· 2025-12-22 06:31
LLM Development & Training - The report introduces a method to build a modern LLM from scratch using Karpathy's nanochat, emphasizing its clean, minimal, and hackable codebase [1] - The process involves training a tokenizer, pre-training for next-word prediction, mid-training for conversational abilities, and SFT (fine-tuning) on high-quality dialogue datasets [1] - Evaluation and logging are integral to every step of the LLM development process [1] Implementation & Accessibility - The method can be reproduced with a single click on a LightningAI studio, requiring zero setup [1]
X @Avi Chawla
Avi Chawla· 2025-10-23 20:02
Core Concept of Memento - Memento reframes continual learning as memory-based online reinforcement learning over a memory-augmented MDP, learning from experiences using memory instead of updating LLM weights [2] - Memento aims to improve AI agent performance from experience without fine-tuning LLM weights [1] Key Components - Case-Based Reasoning (CBR) decomposes complex tasks into sub-tasks and retrieves relevant past experiences [2] - Executor executes each subtask using MCP tools and records outcomes in memory for future reference [3] MCP Tools - MCP tools enable the executor to accomplish most real-world tasks [3] - MCP tools include Web research, Document handling, Safe Python execution, Data analysis, and Media processing [3]
X @Avi Chawla
Avi Chawla· 2025-08-11 06:31
Model Fine-tuning - Fine-tuning enables the LLM to generate reasoning tokens in French before the final English response [1] - The video demonstrates the LLM's behavior before and after fine-tuning [1]
The 2025 AI Engineering Report — Barr Yaron, Amplify
AI Engineer· 2025-08-01 22:51
AI Engineering Landscape - The AI engineering community is broad, technical, and growing, with the "AI Engineer" title expected to gain more ground [5] - Many seasoned software developers are AI newcomers, with nearly half of those with 10+ years of experience having worked with AI for three years or less [7] LLM Usage and Customization - Over half of respondents are using LLMs for both internal and external use cases, with OpenAI models dominating external, customer-facing applications [8] - LLM users are leveraging them across multiple use cases, with 94% using them for at least two and 82% for at least three [9] - Retrieval-Augmented Generation (RAG) is the most popular customization method, with 70% of respondents using it [10] - Parameter-efficient fine-tuning methods like LoRA/Q-LoRA are strongly preferred, mentioned by 40% of fine-tuners [12] Model and Prompt Management - Over 50% of respondents are updating their models at least monthly, with 17% doing so weekly [14] - 70% of respondents are updating prompts at least monthly, and 10% are doing so daily [14] - A significant 31% of respondents lack any system for managing their prompts [15] Multimodal AI and Agents - Image, video, and audio usage lag text usage significantly, indicating a "multimodal production gap" [16][17] - Audio has the highest intent to adopt among those not currently using it, with 37% planning to eventually adopt audio [18] - While 80% of respondents say LLMs are working well, less than 20% say the same about agents [20] Monitoring and Evaluation - Most respondents use multiple methods to monitor their AI systems, with 60% using standard observability and over 50% relying on offline evaluation [22] - Human review remains the most popular method for evaluating model and system accuracy and quality [23] - 65% of respondents are using a dedicated vector database [24] Industry Outlook - The mean guess for the percentage of the US Gen Z population that will have AI girlfriends/boyfriends is 26% [27] - Evaluation is the number one most painful thing about AI engineering today [28]
How to build Enterprise Aware Agents - Chau Tran, Glean
AI Engineer· 2025-07-24 09:22
[Music] Thanks Alex for the introduction. That was a very impressive LLM generated summary of me. Uh I've never heard it before but uh nice.Um so um today I'm going to talk to you about something that has been keeping me up at night. Uh probably some of you too. So how to build enterprise aware agents.How to bring the brilliance of AI into the messy complex realities of uh how your business operated. So let's jump straight to the hottest question of the month for AI builders. Uh should I build workflows or ...
X @Avi Chawla
Avi Chawla· 2025-07-21 06:40
LLM Training Stages - LLM 从零开始训练包含四个阶段 [1] - 第一步是使用随机初始化的模型 [2] - 之后在大规模语料库上进行预训练 [2] - 使用指令微调使其能够遵循命令 [2] - 使用偏好和推理微调来优化响应 [2]
X @Avi Chawla
Avi Chawla· 2025-06-28 21:05
Machine Learning Paradigms - Four machine learning training paradigms are visually explained [1] - The paradigms include Transfer Learning, Fine-tuning, Multi-task Learning, and Federated Learning [1]
X @Avi Chawla
Avi Chawla· 2025-06-28 06:31
4 machine learning training paradigms, explained visually:- Transfer Learning- Fine-tuning- Multi-task Learning- Federated Learning https://t.co/q6wnDTQtIn ...