Workflow
确定性LLM推理
icon
Search documents
超越ZIP的无损压缩来了!华盛顿大学让大模型成为无损文本压缩器
量子位· 2025-10-11 04:09
Core Insights - The article discusses the challenges of data storage arising from the generation of massive data by large language models (LLMs) and introduces an innovative solution called LLMc, which utilizes LLMs for lossless text compression [1][2]. Group 1: LLMc Overview - LLMc has demonstrated superior compression rates compared to traditional compression tools like ZIP and LZMA across various datasets, including Wikipedia, novels, and scientific abstracts [2]. - The project has been open-sourced, with the main author being Yi Pan, an undergraduate from Shanghai Jiao Tong University currently interning at the University of Washington [4]. Group 2: Compression Mechanism - The inspiration for LLMc arose from a challenge related to the non-deterministic nature of LLM inference, which complicated precise and reproducible compression and decompression [5]. - The connection between LLMs and data compression is rooted in Shannon's source coding theorem, which states that the optimal encoding length of a symbol is proportional to its negative log-likelihood [6]. - LLMs, as powerful probability prediction engines, assign high probabilities to the next token in a sequence, enabling efficient compression by transforming high-dimensional distributions into structured probability information [7]. Group 3: Rank-Based Encoding - LLMc employs a clever method known as "rank-based encoding," where instead of storing the token itself, it stores the rank of the token in a predicted probability distribution list [8][10]. - During decompression, the same LLM and context are used to recreate the probability distribution, allowing the system to accurately select the corresponding token based on its stored rank [10][11]. Group 4: Challenges and Limitations - The research team identified several challenges with the current version of LLMc, including computational complexity, which scales quadratically with sequence length, and memory bandwidth limitations for long sequences [12]. - LLMc's processing speed is currently significantly lower than traditional compression algorithms due to its reliance on large-scale model inference [13]. - The implementation is primarily focused on natural language, and future exploration is needed to extend its application to other modalities such as images, videos, or binary data [14].