ShuffleNet

Search documents
「走出新手村」十次 CV 论文会议投稿的经验总结
自动驾驶之心· 2025-06-30 12:33
Core Insights - The article provides a comprehensive guide for newcomers on how to improve the quality and acceptance rate of research papers in the field of deep learning, based on the author's personal experiences and reflections during the submission process [2][3]. Paper Production and Submission Process - The typical process for producing and submitting a deep learning paper involves generating a good idea or experimental results, expanding on them, and writing a structured paper according to the conference's requirements [3][4]. - After submission, if there are no serious issues, the paper enters the review stage, where feedback is provided by three reviewers, and authors must respond to comments, often leading to a significant number of papers being withdrawn from consideration [4][5]. Importance of Writing Quality - Writing a good paper is crucial as it serves as a vehicle for conveying ideas and can significantly impact an author's career; high-quality papers are more likely to be cited and recognized [7][8]. - The quality of a paper can reflect an author's research achievements, with a few outstanding papers often defining a scholar's career [7]. Innovation and Core Ideas - The concept of novelty is central to deep learning papers, where innovation can be measured by the impact of the problem addressed, the effectiveness of the solution, and the novelty of the methods used [10][11]. - Authors should clearly define their core ideas and potential impact when selecting topics and writing papers, ensuring that their contributions are well-articulated [11]. Writing Techniques - Effective writing in deep learning papers often follows a structured approach, where the title and abstract are critical for attracting readers and matching appropriate reviewers [13][14]. - The introduction should clearly present the importance of the problem and the proposed solution, while the experimental section should demonstrate the effectiveness of the approach [15][16]. Common Reviewer Feedback - Common negative feedback from reviewers includes perceived lack of understanding of the field, unclear contributions, and failure to respect prior work [22][24]. - Authors are encouraged to address potential issues before submission by considering common criticisms and ensuring their papers are well-structured and clearly articulated [22][24].
专访张祥雨:多模态推理和自主学习是未来的 2 个 「GPT-4」 时刻
海外独角兽· 2025-06-08 04:51
本期内容是拾象 CEO 李广密对大模型公司阶跃星辰首席科学家张祥雨的访谈。 张祥雨专注于多模态领域,他提出了 DreamLLM 多模态大模型框架,这是业内最早的图文生成理解 一体化的多模态大模型架构之一,基于这个框架,阶跃星辰发布了中国首个千亿参数原生多模态大 模型 Step-1V。此外,他的学术影响力相当突出,论文总引用量已经超过了 37 万次。 一直以来,业界都相当期待一个理解、生成一体化的多模态,但直到今天这个模型还没出现,如何 才能达到多模态领域的 GPT-4 时刻?这一期对谈中,祥雨结合自己在多模态领域的研究和实践历 程,从纯粹的技术视角下分享了自己对多模态领域关键问题的全新思考,在他看来,虽然语言模型 领域的进步极快,但多模态生成和理解的难度被低估了: • 接下来 2-3 年,多模态领域会有两个 GPT-4 时刻:多模态推理和自主学习; • o1 范式的技术本质在于激发出 Meta CoT 思维链:允许模型在关键节点反悔、重试、选择不同分 支,使推理过程从单线变为图状结构。 目录 01 研究主线: 重新回归大模型 • 多模态生成理解一体化难以实现的原因在于,语言对视觉的控制能力弱,图文对齐不精确, ...