Workflow
AI版权
icon
Search documents
Sora2生成已故名人视频引亲属不满,OpenAI面临版权麻烦
21世纪经济报道· 2025-10-11 12:25
Core Viewpoint - The article discusses the ethical and copyright issues surrounding AI-generated videos of deceased celebrities, particularly focusing on the case of Robin Williams and the implications of OpenAI's Sora 2.0 release, which has sparked significant controversy and backlash from family members and industry stakeholders [1][2][3]. Group 1: AI Video Generation and Controversy - The release of Sora 2.0 has led to a surge in AI-generated videos featuring Robin Williams, raising concerns about the manipulation of his image and voice without consent [1][3][5]. - Robin Williams' daughter has publicly condemned the creation of AI videos of her father, emphasizing the emotional distress it causes to the family and the disrespect it shows to his legacy [5][6]. - The rapid adoption of Sora 2.0, which reportedly surpassed one million downloads within five days, highlights the growing demand for AI-generated content, but also the challenges of regulating its use [5][6]. Group 2: Legal and Ethical Implications - The article outlines the legal framework in China regarding the posthumous rights of deceased individuals, indicating that family members can claim rights over the deceased's image and voice, which complicates the use of AI in recreating these figures [8][9]. - OpenAI has faced pressure from various stakeholders, including Hollywood unions and family members, to establish clearer boundaries regarding the use of deceased individuals' likenesses in AI-generated content [13][14]. - OpenAI has adjusted its copyright policy from an opt-out to an opt-in mechanism, allowing public figures to control the use of their likenesses in Sora-generated videos, although this does not address the rights of deceased individuals [14][15]. Group 3: Industry Response and Future Directions - The article notes that the backlash against AI-generated content is not isolated, as other companies in the industry have faced similar legal challenges and public outcry regarding copyright infringement [13][16]. - There is a call for a more structured approach to the ethical use of AI in recreating public figures, with suggestions for obtaining explicit consent from deceased individuals' estates and establishing clearer guidelines for AI platforms [9][16]. - The ongoing debate highlights the tension between artistic expression and the rights of individuals, suggesting that the industry is still in the process of finding a balance between innovation and ethical responsibility [16].
Sora的“阳谋”:用分钱模式,破解AI版权的死结
Hu Xiu· 2025-10-11 09:58
Core Insights - OpenAI launched its most powerful video generation model, Sora 2.0, which achieved over one million downloads within five days, surpassing the initial speed of ChatGPT [1] - The rapid adoption of Sora 2.0 has reignited long-standing concerns regarding AI copyright issues, particularly as users began generating fan videos using well-known intellectual properties (IPs) [2][3] - In response to the backlash, major Hollywood agencies and companies like Disney are pressuring OpenAI to take responsibility for copyright infringement, leading to a strategic shift in Sora's operational policies [3][4] Legal Context - The controversy surrounding Sora 2.0 stems from the "opt-out" mechanism that allowed the generation of copyrighted content unless explicitly requested to be removed by the copyright holders, which has been criticized for potentially leading to systemic infringement [4][8] - OpenAI's new "opt-in" policy, announced by CEO Sam Altman, aims to establish a revenue-sharing model with copyright holders, marking a significant shift in the relationship between AI companies and IP owners [4][21] - The legal challenges faced by AI companies include the legitimacy of using copyrighted works for training AI models and the risk of generating content that closely resembles existing copyrighted works [9][10][13] Business Model Implications - The new revenue-sharing model proposed by OpenAI seeks to redefine user-generated content as interactive fan creations, providing copyright holders with more control over their IPs and potential revenue streams [18][19] - This model is compared to YouTube's copyright revenue-sharing system, which could incentivize more creative content while offering copyright holders new monetization opportunities [19][22] - However, the implementation of this model faces challenges, including the complexity of tracking and attributing copyright elements in generated content, as well as the need for a clear and fair pricing structure for IP licensing [20][23] Industry Outlook - The shift from litigation to collaboration between AI companies and copyright holders reflects a broader trend in the industry, where the focus is on finding mutually beneficial solutions to copyright disputes [5][21] - The ongoing debate over AI-generated content and copyright distribution highlights the need for updated legal frameworks and standards to address the unique challenges posed by generative AI technologies [22][23] - OpenAI's approach signals a potential transition for the AI industry from unregulated growth to a more structured licensing phase, emphasizing the importance of innovative institutional designs to navigate the complexities of copyright in the AI era [23]
MiniMax“版权劫”:好莱坞重拳下负重前行?
3 6 Ke· 2025-10-08 00:41
Core Viewpoint - The lawsuit filed by major Hollywood studios against Chinese AI company MiniMax highlights the ongoing conflict between technology and copyright, as MiniMax is accused of creating a "piracy business model" by using protected characters to train its AI system and generate unauthorized content [1][2]. Group 1: Lawsuit Details - The lawsuit was filed on September 16, 2025, by Disney, Universal Pictures, and Warner Bros, accusing MiniMax of systematically copying valuable copyrighted characters to profit from unauthorized videos [1][2]. - MiniMax's users can generate high-quality videos featuring iconic characters like Spider-Man and Darth Vader by simply inputting prompts, which has raised significant concerns among the studios [2]. - The studios had previously sent cease-and-desist letters to MiniMax, but the company did not respond substantively, prompting the legal action [2]. Group 2: MiniMax's Global Expansion - Prior to the lawsuit, MiniMax had rapidly expanded internationally, claiming over 157 million users across more than 200 countries [3]. - The company launched several AI applications, including Glow and Talkie, with Talkie achieving over 11 million monthly active users by 2024, primarily from the U.S. market [3]. - MiniMax's revenue was projected to exceed $70 million in 2024, with Talkie being a significant contributor to this growth [3]. Group 3: Financing and Market Position - MiniMax has been a capital darling, nearing completion of a $300 million funding round in 2025, which would elevate its valuation to over $4 billion [4]. - The company previously raised $600 million in March 2024, with Alibaba leading the round, and had a valuation of $2.5 billion at that time [4]. Group 4: Implications of the Lawsuit - The lawsuit poses a significant threat to MiniMax's financing and IPO plans, with potential damages reaching hundreds of millions of dollars, which would be a substantial burden for a company with annual revenue of around $70 million [5][6]. - The plaintiffs are seeking either profits gained from the infringement or statutory damages of up to $150,000 per infringed work under U.S. copyright law [5]. Group 5: Future Outlook - Legal experts suggest that the outcome of the case may hinge on the ongoing debate regarding whether AI training constitutes fair use, which remains a contentious issue in U.S. courts [7]. - A recent settlement involving Anthropic, which paid $1.5 billion to resolve a similar lawsuit, may serve as a reference for MiniMax, although the company faces unique challenges as a rapidly expanding Chinese AI firm in the U.S. market [8][9]. - The case could have broader implications beyond legal matters, reflecting the complexities of U.S.-China tech competition and the evolving landscape of copyright in the age of AI [9].
Meta否认用相册图训练AI|南财合规周报(第196期)
Group 1: AI and Copyright Developments - Meta has stated that it does not currently use unpublished user photos to train its AI models, despite previous reports suggesting otherwise [2][3] - Recent U.S. court rulings have determined that using published works to train AI models can fall under "fair use," with two cases involving Anthropic and Meta providing legal clarity [4][5] - The rulings indicate that while AI companies may have some leeway, the methods of data collection must still be scrutinized to avoid infringement [4] Group 2: Regulatory Changes - The revised Anti-Unfair Competition Law in China will take effect on October 15, 2025, aiming to curb "involution" in competition and establish fair competition review systems [5][6] - The new law prohibits large enterprises from abusing their dominant positions to delay payments to small and medium-sized enterprises [6] - The law also restricts platform operators from forcing other businesses to comply with pricing rules that disrupt market order [6] Group 3: Personal Information Protection - The National Cybersecurity Center reported that 45 mobile applications were found to be illegally collecting and using personal information without user consent [7] - In Shanghai, authorities are actively addressing the misuse of AI technologies, particularly focusing on the protection of personal information rights [8] - Ongoing enforcement actions aim to combat AI misuse, including the generation of inappropriate content and violations of personal data rights [8]
AI版权关键进展:美国连判两案,大模型“偷书”不算偷
Group 1 - The core issue revolves around whether using human works to train AI without authorization constitutes copyright infringement, with recent U.S. court rulings providing new references for this ongoing debate [1][2] - The rulings from the U.S. Northern District Court of California found that both Anthropic and Meta's use of copyrighted works for AI training fell under the "fair use" doctrine, emphasizing that the purpose of use was transformative and did not directly replace the original works [2][3] - The court highlighted that the determination of "fair use" is nuanced and depends on the legality of the data acquisition methods, with a distinction made between legal and illegal sources [4][5] Group 2 - In the Meta case, the court noted that the AI training was for a highly transformative purpose, as it was not intended for reading or disseminating the original works, but rather for generating tasks like writing code or emails [2][3] - The court also emphasized the importance of market impact, stating that if AI outputs could harm the market for original works, it might not qualify as fair use, although this was not proven in the Meta case [7][8] - The Anthropic case similarly recognized the transformative nature of AI training but differentiated between legal and illegal data sources, ruling that using data from illegal sources like "shadow libraries" constituted infringement [6][7] Group 3 - The rulings indicate a cautious approach by the courts, as they do not grant AI companies a blanket permission to use copyrighted works for training, stressing that each case must be evaluated on its own merits [3][6] - The distinction between the two cases lies in the treatment of data sources, with Meta's use of "shadow libraries" being viewed more leniently due to its failed attempts to obtain licenses, while Anthropic's establishment of a permanent internal library from illegally sourced materials was deemed infringing [5][7] - The ongoing legal disputes extend beyond literature, with similar copyright issues emerging in the film and visual arts sectors, indicating a broader industry concern regarding AI training practices [8]
一场就得“数百万美金”?Getty CEO说:“AI版权战”太贵了!
Sou Hu Cai Jing· 2025-05-29 02:46
Core Viewpoint - Getty Images has positioned itself as a staunch defender of artists' rights in the ongoing AI copyright disputes, emphasizing the high costs associated with litigation against AI companies [2][3] Group 1: Getty Images' Actions and Statements - Getty Images banned users from uploading AI-generated images in 2022 and later launched a socially responsible image generator while suing an AI company for not compensating artists [2] - CEO Craig Peters revealed that Getty has spent "millions of dollars" on a copyright lawsuit against Stability AI, highlighting the prohibitive costs of pursuing every infringement case [2] - Getty filed a lawsuit against Stability AI in 2023, claiming that the company used over 12 million images from Getty's library without permission to train its model [2][3] Group 2: AI Companies' Defense and Industry Implications - AI companies argue that their practices of scraping images for model training fall under "fair use," which is protected by copyright law [3] - Stability AI and other AI firms assert that requiring them to pay licensing fees would hinder technological innovation and the growth of the AI industry [5] - Peters criticized this stance, arguing that rights holders should not bear the high costs of litigation against claims that paying artists would stifle innovation [5] Group 3: Public Reactions and Broader Context - The comments from Peters coincided with backlash against Nick Clegg, former Meta global affairs head, for reiterating the AI industry's argument that requiring artist consent would harm the sector [5][6] - Critics have drawn parallels between the current arguments of AI companies and past defenses used by illegal file-sharing platforms like Napster [6] - Getty has submitted recommendations to the Trump administration, urging it to reject AI companies' proposals for exemptions that would allow them to avoid paying artists for their work [7][8]
速递|用8000万授权数据挑战Midjourney,Freepik的生成式AI版权新解法
Z Potentials· 2025-04-30 04:25
图片来源: Freepik 在线平面设计平台 Freepik 于周二发布了一款新型"开放" AI 图像模型,该公司称该模型仅基于商业授权、"适合工作环境"的图片进行训练。 该模型名为 F Lite ,包含约 100 亿个参数——参数是构成模型的内部组件。 据 Freepik 透露, F Lite 是与 AI 初创公司 Fal.ai 合作开发,并利用 64 台 Nvidia H100 GPU 耗时两个月训练完成。 F Lite 加入了基于授权数据训练的小型但不断增长的生成式 AI 模型行列。 推特原文:我们已秘密研发数月!终于能分享它,感觉太棒了! • 常规版:更可预测且忠于提示,但艺术性较低: https://t.co/MyWsKer9Ir • 纹理版:更为混乱且易出错,但能呈现更佳的纹理效 pic.twitter.com/GX5mIpYE8O (@javilopen) 2025 年 4 月 29 日 生成式 AI 正成为针对 OpenAI 和 Midjourney 等 AI 公司的版权诉讼核心。 这类技术常利用来自网络公开渠道的海量内容(包括受版权保护的材料)进行开发。多数开发此类模型的公司主张合理使用原则 ...