
"Token New Era" – A "Ten Questions, Ten Answers" on China's AI Industry
J.P. Morgan systematically answers ten core questions about the AI industry. China's AI sector is shifting from a price war to a focus on model quality. In agent scenarios, 'task completion rate' is far more critical than 'token price' for customer retention. The core profitability challenge is whether gross profit growth can consistently outpace R&D expenditure growth. Both Zhipu and MiniMax are projected to achieve profitability from 2029 onwards, and the overall market will not be a winner-take-all scenario
China's foundational AI model industry is transitioning from an "expectation-driven" to a "demand-driven" phase. In a recent research report, J.P. Morgan systematically addressed ten core questions from investors regarding this sector, asserting that model quality has become the primary determinant of market structure, and industry divergence will accelerate.
According to the report released by J.P. Morgan on March 27, the Chinese AI market is at a distinct inflection point. Demand growth in coding and agent scenarios is accelerating. Domestic model capabilities have reached or surpassed the level of leading U.S. models from a year ago, while local pricing is more economically viable, collectively improving the returns on deployment.
2026 will be a pivotal year for Chinese enterprises' AI demand to potentially replicate the growth trajectory seen in the U.S. in 2025. Using Anthropic as a benchmark, its Annual Recurring Revenue (ARR) surged from $1 billion in December 2024 to $19 billion in March 2026, an approximate 19-fold increase within 15 months.
The Chinese market possesses the conditions to follow a similar path, especially in the coding domain. Internet giants such as Tencent, Alibaba, and ByteDance have integrated relevant tools into their existing ecosystems, driving demand from standalone demonstrations to comprehensive deployment. The firm maintained its "Overweight" rating on Zhipu and MiniMax, with target prices of HKD 800 and HKD 1100 respectively.
Question 1: Is AI demand linear or explosive at an inflection point?
Demand is driven by inflection points, not linear growth.
As long as model quality is sufficient to unlock real-world application scenarios, usage will shift from linear growth to an "upward convex curve" explosion. The most compelling evidence comes from the U.S. market: Anthropic's Annual Recurring Revenue (ARR) soared from $1 billion in December 2024 to $19 billion in March 2026, a nearly 19-fold increase in just 15 months.
China currently possesses the foundational conditions for a similar boom: domestic model capabilities have surpassed those of leading U.S. models from a year ago, and local pricing aligns better with China's economic realities, collectively improving the expected returns on AI implementation.
On the agent side, OpenClaw has become a significant catalyst, pushing use cases from single-turn interactions to multi-step task execution, thereby substantially increasing the token consumption per task. Internet giants like Tencent, Alibaba, and ByteDance have integrated OpenClaw-related tools into their existing ecosystems, signaling a transition from "developer experiments" to "comprehensive ecosystem deployment".

Question 2: Will API pricing rise, fall, or diverge?
Pricing will not move unilaterally; divergence is the main theme.
On one hand, models with superior capabilities command pricing power. If a model can uniquely unlock high-value tasks (e.g., agent coding, long-term workflows, enterprise-grade reliability), customers are willing to pay a premium because the returns are quantifiable. On the other hand, as hardware and algorithmic efficiency continue to improve, the cost per unit of inference will steadily decline, exerting price pressure on models with stagnant capabilities.
The eventual outcome is a differentiated pricing structure: models that consistently maintain cutting-edge capabilities can achieve simultaneous increases in both volume and price; models that fail to iterate continuously will face price declines, and even with growing usage, their profit margins will become uncertain.

Question 3: If pricing isn't the main battlefield, where is the focus of competition?
The main battlefield has shifted from token prices to model capabilities.
This is a key change compared to last year – the focus in the Chinese market in 2025 was a comprehensive price war, whereas now, quality is far more important than unit price in the fastest-growing coding and agent scenarios.
In multi-step workflows, customers are essentially buying "successful task completion," not "cheap tokens." The research report provides an intuitive mathematical example: if the success rate of a single step increases from 85% to 98%, the final completion rate for a 20-step task will jump from 4% to 67%. Under this logic, the model with the lowest price per token might actually have the highest effective comprehensive cost per task.
The report also points out that companies with strong cutting-edge models can easily extend into the low-end market, but companies relying solely on low prices struggle to move upmarket.

Question 4: Why is the foundational large model sector still a "life-and-death struggle" industry?
Small technical gaps, endless iteration cycles, and converging monetization models – these three factors dictate the industry's extreme ruthlessness.
The capability gaps between different large model companies in China are often smaller than investors anticipate, making the market highly unstable. In this industry, "standing still" is not a neutral outcome but signifies a loss of position – companies must continuously invest and iterate to avoid falling behind.
The convergence of business models exacerbates the pressure of elimination. Both revenue growth and profit margins primarily depend on product strength, and switching costs remain low. This means companies that lose technological momentum will rapidly lose their defensive capabilities commercially and financially, and the number of truly reliable companies in the industry will gradually decrease.
Question 5: What determines profitability?
The core issue is whether gross profit growth can consistently outpace the growth of R&D expenditure.
The basic economic model for the token business is clear: Revenue = Token Volume × Price, with the main cost being inference computation and the largest operational expense being training-related R&D. As model efficiency and inference chip efficiency continue to improve, the gross profit margin for cutting-edge models should gradually increase.
However, the outlook for operating profit is more complex. Anthropic serves as a cautionary example: even with monthly revenue reaching $14 billion in February 2026, the company announced a new $30 billion funding round concurrently, emphasizing continuous cutting-edge development – high revenue does not equate to normalized training intensity.
The baseline scenario is that both Zhipu and MiniMax are projected to achieve profitability from 2029 onwards. The report emphasizes that more important tracking metrics than the specific year of profitability are the sustained growth trend in usage volume and the continuous improvement in unit economics.

Question 6: How should investors track model strength?
It requires combining token price, usage volume, and third-party evaluations; a single metric is insufficient.
-
Token Price: This is the most important indicator as it represents the company's real-time expression of its product's market positioning. The price difference with the best models is becoming a good proxy for actual competitive strength.
-
Token Usage Volume: Actual consumption reflects the genuine choices of users and developers. Third-party API aggregators like OpenRouter can serve as references, with particular attention needed for the growth of agent-based workloads, as these consume significantly more tokens per task than simple workflows.
-
Third-Party Evaluations: Artificial Analysis provides structured evaluations, and LMArena reflects the blind preference of real users. These two complement each other to form a more complete external perspective.

Question 7: With internet giants heavily entering the B2B space, what is the future for independent model companies?
Competitive boundaries are converging, ultimately returning to the competition of model capabilities.
Alibaba has clearly prioritized Cloud and AI, deeply integrating model development with enterprise workflows. Tencent's agent products now cover personal, developer, and enterprise scenarios. OpenAI is also shifting its commercial focus to enterprise products and coding deployment. The direction for leading companies is consistent: AI is evolving from a "consumer-side feature" to a "tool that directly generates enterprise revenue."
In this context, independent model companies can no longer rely solely on the "cloud-neutral" label for a moat, nor can internet giants fully compensate for model deficiencies solely through ecosystem traffic advantages. When deploying AI, enterprise customers primarily purchase model quality – stronger coding inference capabilities and more reliable workflow completion rates.
Question 8: What factors determine a company's survival?
Talent first, compute second, organization third – all three are indispensable.
-
Top Research Talent: This remains a research-driven industry. High-level technical judgment is itself a competitive factor; management's ability to make correct decisions on research directions directly impacts the company's technological trajectory.
-
Compute Power and Capital: Cutting-edge training is prohibitively expensive, and inference economics depend on infrastructure quality. Weak compute acquisition capability is a structural disadvantage – it not only affects model training efficiency but also weakens the ability to respond to demand at a reasonable cost.
-
Organizational Execution: In a rapidly iterating market, the ability to translate research findings into products, products into usage, and usage into monetization is almost as important as the model itself.
Question 9: If everyone is progressing, will models eventually converge? Will there be a winner-take-all market?
Overall strength will approach, but not converge; the market will not form a winner-take-all structure.
Different companies have divergences in architectural choices, training data, product focus, and technological paths. These differences will continue to yield distinct capability advantages. The report suggests that in a still rapidly expanding market, multiple companies can grow simultaneously, even with some overlap in capabilities – the expansion of the overall market at this stage is far more significant than prematurely worrying about commoditization.
In the long run, a more realistic market endgame is not "one dominant player and the rest exit," but rather a few truly capable companies remaining, each with its own areas of strength, competing in a market large enough to support multiple winners. As AI expands from productivity tools to consumer scenarios, differences in individual tastes, styles, and preferences will further reinforce this diversified landscape.
Question 10: How to unify understanding of open-source/closed-source, model iteration, and global expansion risks?
Iteration is mandatory, open-source/closed-source are strategic choices, and the core risks of global expansion lie in compute power and compliance.
Regarding model iteration, the expected pace is roughly one flagship model generation per year (e.g., from GLM 4.7 to GLM 5, MiniMax M2 series to M3 series), with minor upgrades driven by reinforcement learning in between. Halting iteration means losing competitive standing.
On open-source vs. closed-source, the report believes the answer is not either/or. Closed-source models offer stronger commercial defensibility and reduce the risk of disintermediation; open-source aids ecosystem building, increases adoption rates, and accelerates technical feedback. Therefore, most Chinese model companies will ultimately adopt a hybrid strategy: keeping the latest and strongest models closed-source while open-sourcing some other versions.
In terms of global expansion, the biggest risk remains compute power acquisition. Both training and inference are highly dependent on high-performance chips, and tightening export controls will simultaneously weaken model progress speed and cost competitiveness. The second risk is data and security compliance: if model deployment, user services, and data storage can be localized overseas, cross-border data transfer issues are relatively manageable; however, local privacy regulations and the determination of data access rights for Chinese affiliated entities remain sources of uncertainty.
