Workflow
open-source
icon
Search documents
X @Avi Chawla
Avi Chawla· 2025-08-10 06:33
Build human-like memory for your Agents (open-source)!Every agentic and RAG system struggles with real-time knowledge updates and fast data retrieval.Zep solves these issues with its continuously evolving and temporally-aware Knowledge Graph.Like humans, Zep organizes an Agent's memories into episodes, extracts entities and their relationships from these episodes, and stores them in a knowledge graph:(refer to the image below as you read)1) Episode Subgraph: Captures raw data with timestamps, retaining ever ...
X @Avi Chawla
Avi Chawla· 2025-08-05 19:33
RT Avi Chawla (@_avichawla)Evaluate conversational LLM apps like ChatGPT in 3 steps (open-source).Unlike single-turn tasks, conversations unfold over multiple messages.This means that the LLM's behavior must be consistent, compliant, and context-aware across turns, not just accurate in one-shot output.In DeepEval, you can do that with just 3 steps:1) Define your multi-turn test case as a ConversationalTestCase.2) Define a metric with ConversationalGEval in plain English.3) Run the evaluation.Done!This will ...
X @Avi Chawla
Avi Chawla· 2025-08-05 06:35
Evaluate conversational LLM apps like ChatGPT in 3 steps (open-source).Unlike single-turn tasks, conversations unfold over multiple messages.This means that the LLM's behavior must be consistent, compliant, and context-aware across turns, not just accurate in one-shot output.In DeepEval, you can do that with just 3 steps:1) Define your multi-turn test case as a ConversationalTestCase.2) Define a metric with ConversationalGEval in plain English.3) Run the evaluation.Done!This will provide a detailed breakdow ...
X @Avi Chawla
Avi Chawla· 2025-08-01 06:30
The missing backend for your AI Agents!Motia is a unified system where APIs, background jobs, events, and agents are just plug-and-play steps.- Python, JS & TypeScript in the same workflow- Built-in observability- One-click deploy100% open-source. https://t.co/ttyF9egrQu ...
X @Avi Chawla
Avi Chawla· 2025-07-31 06:35
Key Features of MCP Server - MongoDB released an open-source MCP Server enabling AI tools to interact directly with MongoDB deployments [1] - The MCP server allows users to write production-grade queries using natural language [1] - It eliminates the need for manual queries or memorizing syntax [1] Functionality and Use Cases - Users can perform tasks like "Show me the most active users," "Create a new database user with read-only access," and "What's the schema for my orders collection?" using natural language [1] - The Agent handles the execution of these tasks [1] Platform Compatibility - The MCP server is compatible with Atlas, Community Edition, and Enterprise Advanced [1]
Introduction to LLM serving with SGLang - Philip Kiely and Yineng Zhang, Baseten
AI Engineer· 2025-07-26 17:45
SGLang Overview - SGLang is an open-source, high-performance serving framework for large language models (LLMs) and large vision models (VLMs) [5] - SGLang supports day zero releases for new models from labs like Quen and DeepSeek, and has a strong open-source community [7] - The project has grown rapidly, from a research paper in December 2023 to nearly 15,000 GitHub stars in 18 months [9] Usage and Adoption - Base 10 uses SGLang as part of its inference stack for various models [8] - SGLang is also used by XAI for their Glock models, inference providers, cloud providers, research labs, universities, and product companies like Koser [8] Performance Optimization - SGLang's performance can be optimized using flags and configuration options, such as CUDA graph settings [20] - Eagle 3, a speculative decoding algorithm, can be used to improve performance by increasing the token acceptance rate [28][42][43] - The default CUDA graph max batch size on L4 GPUs is eight, but it can be adjusted to improve performance [31][36] Community and Contribution - The SGLang community is active and welcomes contributions [7][54] - Developers can get involved by starring the project on GitHub, filing issues, joining the Slack channel, and contributing to the codebase [9][54][55] - The codebase includes the SGLang runtime, a domain-specific front-end language, and a set of optimized kernels [58]
X @Avi Chawla
Avi Chawla· 2025-07-23 06:30
If you found it insightful, reshare it with your network.Find me → @_avichawlaEvery day, I share tutorials and insights on DS, ML, LLMs, and RAGs.Avi Chawla (@_avichawla):Building front-end Agentic apps just got 10x easier (open-source)!If you're building apps where Agents are part of the interface, not just running in the background, AG-UI protocol has become the standard.For context:- MCP connects agents to tools- A2A connects agents to https://t.co/Sg41TjXiiw ...
X @Avi Chawla
Avi Chawla· 2025-07-13 19:40
RT Avi Chawla (@_avichawla)The only MCP server you'll ever need!MindsDB lets you query data from 200+ sources, like Slack, Gmail, social platforms, and more, in both SQL and natural language.A federated query engine that comes with a built-in MCP server.100% open-source with 33k+ stars! https://t.co/wvoPEqP9lI ...
X @Avi Chawla
Avi Chawla· 2025-07-13 06:33
Core Functionality - MindsDB 提供了一个联邦查询引擎,内置 MCP 服务器 [1] - MindsDB 允许使用 SQL 和自然语言查询来自 200 多个来源的数据,例如 Slack、Gmail、社交平台等 [1] Open Source & Community - 该项目是 100% 开源的,拥有超过 33,000 个 star [1]
X @Avi Chawla
Avi Chawla· 2025-07-09 19:29
RT Avi Chawla (@_avichawla)The fastest serving engine for LLMs is here (open-source)!LMCache is an LLM serving engine designed to reduce time-to-first-token and increase throughput, especially under long-context scenarios.It boosts vLLM with 7x faster access to 100x more KV caches.100% open-source! https://t.co/IfyZzdnq4z ...