Workflow
Avi Chawla
icon
Search documents
X @Avi Chawla
Avi Chawla· 2025-12-05 20:31
Code Quality & AI Integration - SonarQube MCP server detects production-grade code quality issues in real-time [1] - AI code generation shifts engineering bottleneck to code review, developers spend 90% of debugging time on AI-generated code [1] - AI reviewers share blind spots with AI generators, lacking proof checking, system behavior validation, and consequence review [1][3] - SonarQube addresses limitations of AI code review by providing enterprise-grade code analysis and instant feedback [1] SonarQube Capabilities - SonarQube processes over 750 billion lines of code daily, identifying various bug patterns [2] - It identifies security vulnerabilities like SQL injection, XSS, and hardcoded secrets [4] - It detects code smells, technical debt, and maintainability issues [4] - It identifies test coverage gaps [4] Implementation - SonarQube MCP server installation is simple and can be added to AI assistant's config [4] - GitHub repository is available [4]
X @Avi Chawla
Avi Chawla· 2025-12-05 13:42
AI Code Generation & Engineering Bottleneck - AI正在以极快的速度生成代码,但工程瓶颈已经从编写转移到审查 [1] - 开发者现在将 90% 的调试时间花在 AI 生成的代码上 [1] MCP Server & Code Quality - MCP服务器可以实时检测生产级代码质量问题 [1]
X @Avi Chawla
Avi Chawla· 2025-12-05 06:31
SonarQube MCP server:(don't forget to star it ⭐)https://t.co/oqcTUdZsWE ...
X @Avi Chawla
Avi Chawla· 2025-12-05 06:31
Core Problem & Solution - AI 代码生成提速,但工程瓶颈转移至代码审查,开发者 90% 的调试时间用于 AI 生成的代码 [1] - AI 代码审查存在盲点,与 AI 代码生成器有相同的根本缺陷 [1] - SonarQube MCP Server 提供企业级代码分析,针对漏洞、代码异味等提供即时反馈 [1] SonarQube Capabilities - SonarQube 每日处理超过 7500 亿行代码,积累了丰富的 bug 模式经验 [2] - SonarQube 检测安全漏洞(SQL 注入、XSS、硬编码密钥等)[4] - SonarQube 识别代码异味和技术债务 [4] - SonarQube 发现测试覆盖率缺口 [4] - SonarQube 评估可维护性问题 [4] AI Reviewer Limitations - AI 审查器进行模式匹配,而非验证 [3] - AI 审查器验证语法,而非系统行为 [3] - AI 审查器审查代码,而非后果 [3] Setup - 安装 SonarQube MCP 服务器 [4] - 将其添加到 AI 助手的配置中 [4]
X @Avi Chawla
Avi Chawla· 2025-12-04 19:38
LLM Fine-tuning Techniques - Traditional fine-tuning is impractical for LLMs due to the large number of parameters (billions) and data size (hundreds of GBs), leading to the development of parameter-efficient fine-tuning (PEFT) [1] - PEFT techniques involve finding a lower-rank adaptation of LLM weight matrices [2] Specific PEFT Techniques - **LoRA (Low-Rank Adaptation):** Adds two low-rank trainable matrices (A and B) alongside weight matrices, adjusting updates in these low-rank matrices instead of fine-tuning the original weights, significantly reducing memory usage [3] - **LoRA-FA (Frozen-A):** Freezes matrix A in LoRA and only updates matrix B, further reducing activation memory requirements [4] - **VeRA:** Freezes matrices A and B, sharing them across all layers, and learns layer-specific scaling vectors instead [4] - **Delta-LoRA:** Tunes the original weight matrix W by adding the difference (delta) between the product of matrices A and B in two consecutive training steps [4][5] - **LoRA+:** Sets a higher learning rate for matrix B compared to matrix A in LoRA, resulting in better convergence [6]
X @Avi Chawla
Avi Chawla· 2025-12-04 06:30
I have been fine-tuning LLMs for over 2 years now!Here are the top 5 LLM fine-tuning techniques, explained with visuals:First of all, what's so different about LLM finetuning?Traditional fine‑tuning is impractical for LLMs (billions of params; 100s GB).Since this kind of compute isn't accessible to everyone, parameter-efficient finetuning (PEFT) came into existence.Before we go into details of each technique, here's some background that will help you better understand these techniques:LLM weights are matric ...
X @Avi Chawla
Avi Chawla· 2025-12-03 19:06
RT Avi Chawla (@_avichawla)Bias-variance tradeoff has a missing detail!Not many ML engineers know about it.Consider fitting a polynomial regression model on some dummy dataset, say, y=sin(x) + noise.As shown in the first plot in the image, as we increase the degree (m):- The training loss will go down to zero.- The test (or validation) loss will decrease and then increase.But notice what happens as we continue to increase the degree (m):↳ Test loss decreases again (shown in the second plot)This is called th ...
X @Avi Chawla
Avi Chawla· 2025-12-03 13:19
If you found it insightful, reshare it with your network.Find me → @_avichawlaEvery day, I share tutorials and insights on DS, ML, LLMs, and RAGs. https://t.co/pFxYUsWxlvAvi Chawla (@_avichawla):Bias-variance tradeoff has a missing detail!Not many ML engineers know about it.Consider fitting a polynomial regression model on some dummy dataset, say, y=sin(x) + noise.As shown in the first plot in the image, as we increase the degree (m):- The training loss will go https://t.co/BIdCfkZRHO ...
X @Avi Chawla
Avi Chawla· 2025-12-03 06:44
Here's the exact time stamp where Ilya Sutskever talks about it on Lex Fridman's podcast: https://t.co/HermsBI3eB https://t.co/3A3EtGpDqx ...
X @Avi Chawla
Avi Chawla· 2025-12-03 06:44
Bias-variance tradeoff has a missing detail!Not many ML engineers know about it.Consider fitting a polynomial regression model on some dummy dataset, say, y=sin(x) + noise.As shown in the first plot in the image, as we increase the degree (m):- The training loss will go down to zero.- The test (or validation) loss will decrease and then increase.But notice what happens as we continue to increase the degree (m):↳ Test loss decreases again (shown in the second plot)This is called the “double descent phenomeno ...