Workflow
大模型评估
icon
Search documents
给AI打个分,结果搞出17亿估值独角兽?
3 6 Ke· 2026-01-07 11:04
Core Insights - LMArena has successfully secured $150 million in Series A funding, raising its valuation to $1.7 billion, marking a strong start to the new year [1] - The funding round was led by Felicis and UC Investments, with participation from Andreessen Horowitz and The House Fund, indicating strong investor confidence in the AI model evaluation sector [3] Company Background - LMArena originated from Chatbot Arena, created by the open-source organization LMSYS, comprised mainly of members from top universities like UC Berkeley and Stanford [4] - The team developed the open-source inference engine SGLang, which achieved performance comparable to DeepSeek's official report on 96 H100 GPUs [4] - The primary focus of LMArena is on evaluating AI models, having established a crowdsourced benchmarking platform during the rise of models like ChatGPT and Claude [6][7] Evaluation Methodology - LMArena employs a unique evaluation method where users anonymously vote on model responses, ensuring unbiased assessments [10] - The platform utilizes an Elo rating system based on the Bradley–Terry model to score models, allowing for real-time updates and fair comparisons [10] - LMArena has become a go-to platform for new models to be tested, with Gemini 3 Pro currently leading the rankings with a score of 1490 [10][11] Growth and Future Plans - Since its seed funding of $100 million last year, LMArena has rapidly expanded, accumulating 50 million votes across various modalities and evaluating over 400 models [12] - The newly raised funds will be used to enhance platform operations, improve user experience, and expand the technical team [12]
给AI打个分,结果搞出17亿估值独角兽???
量子位· 2026-01-07 09:11
Core Insights - LMArena has successfully secured $150 million in Series A funding, raising its valuation to $1.7 billion, marking a strong start to the new year [1][3]. Group 1: Funding and Valuation - The funding round was led by Felicis and UC Investments, with participation from Andreessen Horowitz and The House Fund [3]. - The significant investment reflects the attractiveness of the AI model evaluation sector in the current market [4]. Group 2: Company Background - LMArena originated from Chatbot Arena, which was created by the open-source organization LMSYS following the emergence of ChatGPT in 2023 [5][4]. - The core team consists of highly educated individuals from top universities such as UC Berkeley, Stanford, UCSD, and CMU [6]. Group 3: Technology and Evaluation Methodology - LMArena's open-source inference engine, SGLang, has achieved performance comparable to DeepSeek's official report on 96 H100 GPUs [7]. - SGLang has been widely adopted by major companies including xAI, NVIDIA, AMD, Google Cloud, Oracle Cloud, Alibaba Cloud, Meituan, and Tencent Cloud [8]. - The primary focus of LMArena is on evaluating AI models, which they began with the launch of Chatbot Arena, a crowdsourced benchmarking platform [9][10]. Group 4: Evaluation Process - LMArena employs a unique evaluation process that includes anonymous battles, an Elo-style scoring system, and human-machine collaboration [20]. - Users input questions, and the system randomly matches two models for anonymous responses, allowing users to vote on the quality of the answers without knowing the model identities [21][22]. - The platform's Elo scoring mechanism updates model rankings based on performance, ensuring a fair and objective evaluation process [22]. Group 5: Growth and Future Plans - Since securing $100 million in seed funding, LMArena has rapidly exceeded expectations, accumulating 50 million votes across various modalities and evaluating over 400 models [25]. - The newly raised funds will be used to enhance platform operations, improve user experience, and expand the technical team to support further development [25].
大模型作为评估者的「偏好」困境:UDA实现无监督去偏对齐
机器之心· 2025-11-28 00:51
Core Insights - The article discusses the issue of preference bias in large language models (LLMs) acting as judges, highlighting that even advanced models like GPT-4o and DeepSeek-V3 exhibit systematic favoritism towards their own outputs, leading to significant discrepancies in scoring and ranking [2][4][5] - The introduction of Unsupervised Debiasing Alignment (UDA) offers a new approach to address this bias by allowing models to autonomously adjust scoring rules through unsupervised learning, thus achieving debiasing alignment [2][7] Summary by Sections Problem Statement - Current LLM judging systems, such as Chatbot Arena, face three main challenges: self-preference solidification, heterogeneity bias, and static scoring defects [4][5] - Self-preference solidification leads to models overestimating their own answers, creating a scenario where "who judges wins" [4] - Heterogeneity bias results in varying directions and intensities of bias among different models, ranging from aggressive self-promotion to excessive humility [4] UDA Contribution - UDA transforms the debiasing problem into a sequence learning issue that can be optimized through dynamic calibration, allowing judges to explore optimal scoring strategies autonomously [7][25] - The method utilizes a consensus-driven training approach, treating the collective agreement of judges as a practical optimization target, which helps reduce overall bias [13][18] Methodology - UDA models pairwise evaluations as an instance-level adaptive process, dynamically generating adjustment parameters for each judge model during comparisons [10][11] - The system extracts multiple features from each comparison, including semantic feature vectors and self-perception features, which are crucial for detecting bias tendencies [11][20] Experimental Results - UDA significantly reduces inter-judge variance, lowering the average standard deviation from 158.5 to 64.8, demonstrating its effectiveness in suppressing extreme biases [23] - The average Pearson correlation with human evaluations improved from 0.651 to 0.812, indicating enhanced alignment with human judgment [23] - UDA shows robust zero-shot transfer capabilities, achieving a 63.4% variance reduction on unseen datasets, showcasing its domain-agnostic debiasing ability [23] Conclusion - UDA represents a shift in how judgment calibration is approached, moving away from prompt engineering to a learnable problem, enhancing the robustness and reproducibility of evaluations while aligning more closely with human judgment [25]
3位00后,估值700亿
3 6 Ke· 2025-10-28 12:09
Core Insights - Mercor, an AI recruitment startup, has raised $250 million in new funding, achieving a valuation of $10 billion, which is five times its previous valuation of $2 billion earlier this year [1][3] - Founded in 2023 by three college dropouts, Mercor has developed a large professional talent network and has seen its annual recurring revenue grow from $1 to $500 million in just 17 months [1][3] Company Overview - Mercor specializes in AI-driven recruitment, utilizing AI to screen resumes and match candidates to job positions quickly [3][5] - The company has expanded its services to include data annotation and large model evaluation, leveraging its extensive network of 30,000 experts [3][9] - The startup's revenue has quadrupled since the turmoil at Scale AI, a competitor, leading to an influx of Scale's former employees and clients [13][14] Business Model and Revenue - Mercor's annual recurring revenue reached $70 million by February, driven by its new business in large model evaluation [3][9] - The company manages a network of experts who can earn significant daily wages, with total earnings exceeding $1.5 million daily [9][10] - The new funding will be allocated to expanding the talent network, enhancing the matching system, and improving delivery speed [3][4] Competitive Landscape - Mercor's main competitor, Scale AI, faced challenges after being acquired by Meta, which led to concerns about data neutrality and client trust [13][14] - The controversy surrounding Scale AI has inadvertently benefited Mercor, resulting in a significant increase in its revenue and client base [14][15] Future Prospects - Mercor's AI-driven recruitment model has positioned it as a key player in the large model evaluation space, filling a critical gap in the industry [15][16] - The company aims to continue leveraging its talent network to support the growing demand for high-quality data and expert feedback in AI model development [16]