LLM inference

Search documents
X @Polyhedra
Polyhedra· 2025-08-11 09:34
1/🧵diving deeper into our new work on zkGPT: Proving LLM inference fast with Zero-Knowledge Proofs.Why? Service providers might deploy a smaller/cheaper model than promised. ZK lets them prove correctness without revealing model parameters.📄 https://t.co/ToX51Wla56 ...
X @Avi Chawla
Avi Chawla· 2025-08-06 06:31
1️⃣2️⃣ KV cachingKV caching is a technique used to speed up LLM inference.I have linked my detailed thread below👇 https://t.co/Dt1uH4iniqAvi Chawla (@_avichawla):KV caching in LLMs, clearly explained (with visuals): ...