Workflow
大模型部署
icon
Search documents
联想集团发布企业级"龙虾湖"解决方案,每百万Tokens成本降至1元以下
Ge Long Hui· 2026-03-26 10:24
Core Insights - Lenovo Group has launched the "Lobster Lake" intelligent solution aimed at the enterprise market, integrating various computing platforms and servers to provide a high-performance, private deployment solution for businesses [1] Group 1: Solution Features - The "Lobster Lake" solution offers high intelligence, security, cost-effectiveness, and integrated delivery, with a focus on private deployment [1] - It can achieve a cost of less than 1 yuan per million high-quality tokens, which is one-third of similar cloud service solutions, while maintaining a response latency of under 12 milliseconds [1] - The solution supports up to 1,000 concurrent intelligent agents and can output up to 10 billion tokens daily with a 4K context length [1] Group 2: Addressing Enterprise Challenges - The solution addresses multiple challenges faced by enterprises when transitioning from personal to business applications, including data leakage risks, reasoning errors, context memory loss, insufficient high-concurrency performance, and high operational costs [2] - It features a private deployment architecture that allows enterprises to fully control data flow and access permissions, ensuring data sovereignty [2] Group 3: Security Measures - The "Lobster Lake" solution includes a four-layer security mechanism to protect sensitive industries such as finance and government [2] - Security measures include container-level logical isolation, self-developed core systems with high-intensity security scans, and various protective technologies like WAF and DDoS protection [2] Group 4: Performance Optimization - The solution supports the deployment of large models with parameter scales ranging from 32 billion to 671 billion, allowing for simultaneous operation of multiple models [3] - It features a multi-instance QMD long-term memory system that overcomes single-instance memory limitations, enhancing retrieval accuracy while reducing token consumption [4] Group 5: Deployment Options - The "Lobster Lake" solution offers two deployment modes: centralized and distributed, catering to different enterprise efficiency needs [5] - It supports one-click deployment and uninstallation, along with comprehensive upgrade and maintenance management [5]
长安汽车和复旦大学申请大模型部署相关专利,实现大模型的按需动态调整
Jin Rong Jie· 2025-11-08 06:26
Core Points - Chongqing Changan Automobile Co., Ltd. and Fudan University have applied for a patent titled "Method, Device, Equipment, Storage Medium, and Program Product for Large Model Deployment" with publication number CN120909607A, filed on August 2025 [1] Summary by Categories Patent Application - The patent application involves a method for deploying large models, which includes receiving current operational scenario information and resource usage information from terminals by a cloud server [1] - The method determines whether the current operational scenario and resource usage information are compatible with the specifications of the terminal's current large model [1] - If not compatible, the method utilizes multi-objective constraints and optimization to select a target large model from a model library based on the current operational scenario and resource usage information [1] - The target model's information is then sent to the terminal for deployment, allowing inference through the target large model [1]
即将开课!自动驾驶VLA全栈学习路线图分享~
自动驾驶之心· 2025-10-15 23:33
Core Insights - The focus of academia and industry has shifted towards VLA (Vision-Language Action) in autonomous driving, which provides human-like reasoning capabilities for vehicle decision-making [1][4] - Traditional methods in perception and lane detection have matured, leading to decreased attention in these areas, while VLA is now a critical area for development among major autonomous driving companies [4][6] Summary by Sections Introduction to VLA - VLA is categorized into modular VLA, integrated VLA, and reasoning-enhanced VLA, which are essential for improving the reliability and safety of autonomous driving [1][4] Course Overview - A comprehensive course on autonomous driving VLA has been designed, covering foundational principles to practical applications, including cutting-edge algorithms like CoT, MoE, RAG, and reinforcement learning [6][12] Course Structure - The course consists of six chapters, starting with an introduction to VLA algorithms, followed by foundational algorithms, VLM as an interpreter, modular and integrated VLA, reasoning-enhanced VLA, and a final project [12][20] Chapter Highlights - Chapter 1 provides an overview of VLA algorithms and their development history, along with benchmarks and evaluation metrics [13] - Chapter 2 focuses on the foundational knowledge of Vision, Language, and Action modules, including the deployment of large models [14] - Chapter 3 discusses VLM's role as an interpreter in autonomous driving, covering classic and recent algorithms [15] - Chapter 4 delves into modular and integrated VLA, emphasizing the evolution of language models in planning and control [16] - Chapter 5 explores reasoning-enhanced VLA, introducing new modules for decision-making and action generation [17][19] Learning Outcomes - The course aims to deepen understanding of VLA's current advancements, core algorithms, and applications in projects, benefiting participants in internships and job placements [24]
一万多块,128G内存+96G显存,竟然只要一台小主机就搞定?
Hu Xiu· 2025-09-19 05:00
Core Insights - AMD's Ryzen AI Max+ 395 is highlighted as a powerful tool for deploying large models and generating images, showcasing its capabilities in AI applications [1] Group 1: Product Features - The Ryzen AI Max+ 395 is equipped with 128GB of memory, enabling efficient handling of complex AI tasks [1] - The product is designed to meet the demands of large model deployment, indicating its suitability for advanced AI workloads [1] Group 2: Market Implications - The introduction of the Ryzen AI Max+ 395 positions AMD competitively in the AI hardware market, potentially attracting businesses focused on AI development [1] - The capabilities of the Ryzen AI Max+ 395 may influence industry standards for AI processing power and efficiency [1]
开放几个大模型技术交流群(RAG/Agent/通用大模型等)
自动驾驶之心· 2025-09-04 03:35
Group 1 - The establishment of a Tech communication group focused on large models, inviting participants to discuss topics such as RAG, AI Agents, multimodal large models, and deployment of large models [1] - Interested individuals can join the group by adding a designated WeChat assistant and providing their nickname along with a request to join the large model discussion group [2]