上下文光学压缩
Search documents
DeepSeek-OCR:大模型技术,正站在一个新的十字路口
3 6 Ke· 2025-10-22 23:15
Core Insights - DeepSeek has introduced "DeepSeek-OCR," a model that utilizes "Context Optical Compression," significantly enhancing the efficiency of processing textual information from images [1][2][7] - The model demonstrates that images can serve as efficient carriers of information, challenging the traditional reliance on text-based processing [2][6] Group 1: Image Processing Efficiency - DeepSeek-OCR processes documents by treating text as images, compressing entire pages into a few visual tokens, achieving a tenfold efficiency increase with a 97% accuracy rate [1][2] - Traditional methods require thousands of tokens for a lengthy article, while DeepSeek-OCR only needs about 100 visual tokens, allowing it to handle long documents without resource constraints [2][3] Group 2: System Architecture and Functionality - The system consists of two modules: a powerful DeepEncoder that captures page information and a lightweight text generator that converts visual tokens into readable output [3] - The encoder combines local analysis and global understanding, reducing the initial 4096 tokens to just 256, showcasing a 90% reduction compared to competitors [3][4] - In practical tests, a single A100 GPU can process over 200,000 pages daily, with potential scalability to 33 million pages across multiple servers [3][4] Group 3: Information Density and Model Training - The paradox of image data being more efficient lies in its information density; images can encapsulate more data compactly compared to text tokens, which require extensive dimensional expansion [4][5] - While DeepSeek-OCR proves the feasibility of visual tokens, training purely visual models remains a challenge due to the ambiguity in predicting image segments [5][9] Group 4: Potential Impact and Applications - If widely adopted, this technology could transform the "token economy," significantly reducing processing costs for long documents and enhancing data extraction from complex formats [6][7] - It could also improve chatbots' long-term memory by converting old conversations into low-resolution images, simulating human memory decay while extending context without increasing token consumption [6][11] Group 5: Conclusion - The exploration of DeepSeek-OCR not only achieves a tenfold efficiency improvement but also redefines the boundaries of document processing, challenging existing limitations and optimizing cost structures [7][8]
重磅,DeepSeek再开源:视觉即压缩,100个token干翻7000个
3 6 Ke· 2025-10-21 01:35
一图胜千言!DeepSeek-OCR模型大胆探索视觉-文本压缩边界。通过少量视觉token解码出10倍以上的文本信息,这款端到端VLM架构不仅在 OmniDocBench基准上碾压GOT-OCR2.0,还为LLM的长上下文问题提供高效解决方案。 DeepSeek再发新模型! Github上,DeepSeek新建了DeepSeek-OCR仓库,目的是探索视觉-文本压缩的边界。 常言道:一图胜万言。对LLM也是如此! 在理论上,DeepSeek-OCR模型初步验证了「上下文光学压缩」的可行性—— 从少量视觉token中,模型能够有效解码出超过其数量10倍的文本token。 也就是说,包含文档文本的单张图像,能以远少于等效文本的token量来表征丰富信息。 这表明通过视觉token进行光学压缩可以实现更高的压缩比。 作为连接视觉与语言的中间模态,OCR任务是视觉-文本压缩范式理想的试验场—— 它在视觉与文本表征之间建立了天然的压缩-解压缩映射关系,同时提供可量化的评估指标。 在OCR任务上,DeepSeek-OCR有较高实用价值:在OmniDocBench基准测试中,仅用100个视觉token即超越GOT-OCR2 ...
DeepSeek新模型被硅谷夸疯了!用二维视觉压缩一维文字,单GPU能跑,“谷歌核心机密被开源”
Hua Er Jie Jian Wen· 2025-10-21 00:27
Core Insights - DeepSeek has released an open-source model named DeepSeek-OCR, which is gaining significant attention in Silicon Valley for its innovative approach to processing long texts using visual compression techniques [1][4][21] - The model is designed to tackle the computational challenges associated with large models handling lengthy text, achieving high accuracy rates even with reduced token usage [1][4][5] Model Performance - DeepSeek-OCR operates with a model size of 3 billion parameters and demonstrates a remarkable ability to decode text with high accuracy, achieving 97% accuracy with a compression ratio of less than 10 times and maintaining 60% accuracy even at a 20 times compression ratio [1][4][5] - The model has been benchmarked against existing models, showing superior performance with significantly fewer visual tokens, such as using only 100 visual tokens to outperform models that require 256 tokens [7][8] Data Generation Efficiency - The model can generate over 200,000 pages of high-quality training data daily using a single A100-40G GPU, showcasing its efficiency in data generation [2][4] Innovative Approach - DeepSeek introduces a concept called "Contextual Optical Compression," which compresses textual information into visual formats, allowing the model to interpret content through images rather than text [4][10] - The architecture includes two main components: the DeepEncoder for converting images into compressed visual tokens and the DeepSeek3B-MoE-A570M for reconstructing text from these tokens [10][11] Flexibility and Adaptability - The DeepEncoder is designed to handle various input resolutions and token counts, allowing it to adapt to different compression needs and application scenarios [11][12] - The model supports complex image analyses, including financial reports and scientific diagrams, enhancing its applicability across diverse fields [12][14] Future Implications - The research suggests that this unified approach to visual and textual processing could be a step towards achieving Artificial General Intelligence (AGI) [4][21] - The team behind DeepSeek-OCR is exploring the potential of simulating human memory mechanisms through optical compression, which could lead to more efficient handling of long-term contexts in AI [20][21]
刚刚,DeepSeek重要突破,大模型上下文紧箍咒打破
3 6 Ke· 2025-10-20 23:22
Core Insights - DeepSeek has introduced a novel technology path in the competition of large language models by open-sourcing the DeepSeek-OCR model, which proposes the concept of "Contextual Optical Compression" for efficient information compression through text-to-image conversion [1][8]. Group 1: Model Performance and Capabilities - The feasibility of DeepSeek-OCR has been validated, achieving a decoding accuracy of 97% at a 10x compression ratio, indicating near-lossless compression, while maintaining approximately 60% accuracy at a 20x compression ratio [3][21]. - DeepSeek-OCR can express similar textual content using fewer tokens by converting text tokens into visual tokens, providing a new approach to address the high computational costs associated with processing long texts in large language models [6][11]. - In practical applications, DeepSeek-OCR surpassed GOT-OCR 2.0 using only 100 visual tokens and outperformed MinerU 2.0 with less than 800 visual tokens, demonstrating its efficiency [6][23]. Group 2: Technical Architecture - The architecture of DeepSeek-OCR consists of two main components: DeepEncoder, a visual encoder designed for high compression and high-resolution document processing, and DeepSeek3B-MoE, a lightweight mixture of experts language decoder [12][18]. - DeepEncoder employs a dual-structure design combining local and global attention to achieve high-fidelity visual understanding, significantly reducing the number of vision tokens generated from document images [14][18]. Group 3: Data and Training - DeepSeek-OCR's training process is relatively straightforward, involving independent training of DeepEncoder and the complete DeepSeek-OCR model, utilizing a large dataset for effective learning [20][21]. - The model has been trained on a diverse dataset that includes OCR 1.0 and OCR 2.0 data, general visual data, and pure text data, ensuring robust performance across various document types [25][36]. Group 4: Application and Future Directions - DeepSeek-OCR demonstrates capabilities in deep parsing, allowing it to recognize and extract structured information from various document types, including financial reports and scientific literature [24][29]. - The research team plans to further explore the integration of digital and optical text pre-training methods and evaluate the performance of optical compression in real long-text environments, indicating a promising direction for future research [39].