Author: Zheng Minfang
Source: Wall Street Insights
When OpenAI on the other side of the ocean seems to have pressed the "pause button" on its AI generation model Sora, Chinese tech giants have launched a counterattack in this field.
Recently, ByteDance's latest AI video generation model Seedance 2.0 was launched, and it quickly sparked the internet with advantages such as multimodal input, self-cam movement, and consistency.
Feng Ji, founder of Game Science, after a deep experience, gave a highly significant judgment: "The content field will inevitably experience unprecedented inflation."
Feng Ji's prophecy is not groundless worry.
This shockwave is rapidly spreading to e-commerce, gaming, video platforms, film and television production, and other industries: in the e-commerce sector, the technical barriers of low-end outsourcing and filming bases have been completely flattened; in the gaming industry, the production cycles for concept validation and ad material have been compressed to the extreme, making the competition even more brutal; video platforms are forced to further optimize their distribution logic to cope with the explosive supply growth; and the traditional linear process of "filming + editing" in film production is now facing a dimensional strike from the industrial pipeline of "prompts + generation."
A major industry reshuffle regarding benefits and substitution has already begun.
The explosion of video production capacity
The biggest pain point of AI video in the past year was deliverability.
Whether it's Sora, Runway, domestic Keling, or ByteDance's self-developed Jiemeng, such problems exist. Creators are often trapped in a "gacha" game, needing to generate dozens of times repeatedly to get a few seconds of video that doesn't collapse and maintains consistency.
The core breakthrough of Seedance 2.0 lies in attempting to transform "showmanship" into "deliverable narratives."
The breakthrough of key capabilities is mainly reflected in three aspects:
First is multimodal input. According to AllTech's hands-on testing, a member user who logs in for the first time to JiMeng only needs to pay 1 yuan to enable auto-renewal and can directly use Seedance 2.0, which supports text, images, videos, and audio as reference materials for input. It can be said that almost all formats you can think of can be input to generate videos.
Second, understanding the narrative and learning to operate the camera by oneself. Seedance 2.0 demonstrates "director-level" thinking, capable of not only understanding complex narrative logic but also automatically orchestrating cinematography, completing camera movements such as pushing, pulling, panning, and moving. The video is no longer a simple displacement of static images, but instead possesses cinematic narrative logic.
Third is the consistency of the visuals. According to All-Weather Technology's real testing of multiple AI video generation applications on the market, problems such as facial expressions breaking down during the movement of the main subject and the background showing a mix of clear and blurry areas frequently occur.
But from the demo video, Seedance 2.0 maintains consistency in facial features, visuals, and other information during the main body's movement process, making coherent narrative expression possible.
This means that AI video generation is evolving from a toy into a tool. The ability to turn video generation into a standardized industrial pipeline makes the slogan "everyone is a director" no longer an empty phrase, and it will also significantly reduce the cost of video production.
Feng Ji used "inflation" to describe this change.
"The production cost of general video will no longer be able to follow the traditional logic of the film and television industry, but will gradually approach the marginal cost of computing power. The content field will inevitably experience unprecedented inflation, and traditional organizational structures and production processes will be completely restructured. I believe that anyone who has used it can quickly understand that this prediction is by no means a groundless worry," Feng Ji stated.
First wave of impact
When the marginal cost of video production approaches zero, business models built upon traditional cost structures will be hit first.
The four industries of e-commerce, gaming, video platforms, and film and television production may be the first wave of industries affected.
The most direct impact is first felt in the e-commerce sector.
Product displays, scene performances, and function explanation videos are essentially not dependent on complex artistic narratives, but on clear information delivery.
With the popularity of Seedance 2.0, the threshold for merchants to access video expression capabilities has been completely flattened. Low-end video outsourcing companies and Taobao shooting bases that previously relied on "information gaps" and "technical barriers" to survive will face a harsh winter. Video production may shift from professional outsourcing services to in-house daily operations for merchants.
Compared to e-commerce, the impact of AI video generation models on gaming may still be relatively limited, but the revolution has already quietly begun.
The costs of worldviews deduction, concept validation, and user acquisition material videos are decreasing exponentially. More projects will be validated at earlier stages and also be eliminated at earlier stages.
An insider from a Beijing-based gaming company told The Tech Journal that the company has started small-scale testing for Seedance 2.0 internally.
AI video generation models are also changing the distribution logic of video platforms.
For platforms like Douyin and Kuaishou, videos generated by models such as Seedance 2.0 bring an explosive surge in content supply, forcing the core competitiveness of the platforms to completely shift to the "filtering and distribution" mechanism. For example, whose algorithm can more accurately sift out the gold from the massive amount of AI-generated content, and whose commercial conversion efficiency is higher, will be the winner.
In the film and television industry, Seedance 2.0's multi-camera storytelling capabilities could reshape production workflows.
The birth of a past film or television work often followed a strict linear industrial process: first, a large amount of material was shot, and then editors made selections and spliced them together in the editing room to construct the narrative logic.
But in the logic of Seedance 2.0, this boundary is becoming blurred.
In the filming stage, there is a possibility that future sets could be generated at low cost by AI models; the models themselves have an understanding of camera movement and narrative pacing, and in generating the video, they actually simultaneously complete the "editing" work.
AI is no longer just spitting out fragmented raw footage, but instead directly delivering finished videos with coherent spatiotemporal relationships.
This means that the time-consuming post-production editing phase in traditional film and television production is facing the risk of being "dimensionally reduced" by algorithms.
Future production workflows may no longer be "shooting + editing," but rather "prompts + generation," with editors' roles transforming from "operators" to "instruction engineers" or "aesthetic gatekeepers."
Although the videos generated by the current Seedance 2.0 are not yet perfect, there is still room for improvement in logical details, visuals, etc., these problems will not be obstacles in the near future, given the speed of technological iteration far exceeds market expectations.
The "moat" of IP
Seedance 2.0's amazing "remastering" capability brings the joy of creation to ordinary people, but also puts unprecedented pressure on copyright holders.
Recently, a large number of "remix" and even "parody" clips of Stephen Chow's classic movies have been widely spread on short video platforms.
With the computational power of AI video generation models, a large number of users have been able to replicate Zhou Xingchi's facial expressions, signature laughter, and classic dialogue style at low cost, even generating many absurd plots that never actually happened.
This quickly drew the attention of Stephen Chow's team.
Chen Zhenyu, the agent of Stephen Chow, publicly questioned: "I would like to ask, does this constitute infringement (especially the large amount of spreading in the past two days)? I believe the creators have already profited, and is a certain platform just letting it go and providing it for users to generate and post?"
This challenge seems to reveal copyright anxieties in the AI era, but from a deeper commercial logic perspective, it actually demonstrates the extreme scarcity of top-tier intellectual property in the AI era.
In the future, amid the flood of AI-generated content, the technology itself will no longer be a barrier, as everyone will have the same Seedance 2.0 tool.
The real barriers still lie in the hands of IP owners.
It is precisely because the market is flooded with a large number of "high仿" (high imitation) Zhou Xingchi that the irreplaceability of the "Real·Zhou Xingchi" IP becomes more evident.
When content supply is not only excessive but also "inflated," users' time and attention will become more valuable than ever. What can instantly capture users' attention are still those classic IPs that have been time-tested and possess strong emotional penetration.
In other words, although AI lowers the threshold for production, it infinitely elevates the value of "recognizability."
The outlook remains bright for IP owners. IP assets accumulated over the years will no longer be merely targets of infringement, but can achieve exponential commercial value amplification through legitimate licensing, leveraging AI and the hands of countless creators.
From the launch of Sora 1.0 by OpenAI in February 2024, which became the world's first AI video generation model capable of producing up to 60 seconds of video, to today's Seedance 2.0 by ByteDance, which achieves multimodal input to generate 60-second native audio narrative films, it has only taken 2 years.
In this era of rapid technological development, every industry is standing at a crossroads: the cost at the execution level is being compressed infinitely, and those repetitive, labor-intensive, and time-consuming jobs will be replaced ruthlessly; at the same time, the value of IP and creativity is being amplified infinitely.
When tools become readily available, what determines the quality of content will no longer be whether one knows how to use the software, but whether the vision one has in their mind about the world is sufficiently unique.
