Ginlix AI
50% OFF

In-Depth Analysis of Zhipu AI: R&D Investment, GLM Architecture, and High Gross Margin

#ai #llm #zhipu_ai #gross_margin #r_and_d_investment #glm_architecture #maas #moe_architecture
Neutral
A-Share
January 15, 2026

Unlock More Features

Login to access AI-powered analysis, deep research reports and more advanced features

In-Depth Analysis of Zhipu AI: R&D Investment, GLM Architecture, and High Gross Margin

About us: Ginlix AI is the AI Investment Copilot powered by real data, bridging advanced AI with professional financial databases to provide verifiable, truth-based answers. Please use the chat box below to ask any financial question.

Based on the latest information I have collected, I will now conduct a systematic and comprehensive analysis for you.


In-Depth Analysis of Zhipu AI: Supported by R&D Investment, GLM Architecture, and High Gross Margin
I. Overview of Core Financial Data

According to Zhipu AI’s Hong Kong IPO prospectus and publicly disclosed data [1], the company’s financial performance exhibits the typical characteristics of a large model startup: ‘high growth, high investment, high losses’:

Financial Indicators 2022 2023 2024 H1 2025
Revenue (CNH 100 million)
0.57 1.25 3.12 1.91
R&D Expenses (CNH 100 million)
0.84 5.29 21.95 15.95
R&D Expense Ratio
147% 425%
703%
835%
Gross Margin
54.6% 64.6% 56.3% 50.0%
Net Loss (CNH 100 million)
1.44 7.88 29.58 23.58

As can be seen from the data, the R&D expense ratio reached 703% in 2024, meaning R&D investment was 7 times revenue; this ratio further climbed to 835% in H1 2025 [2]. This ratio is significantly higher than that of international competitors — OpenAI’s R&D expense ratio for the same period was approximately 1.56:1, and Anthropic’s was about 1.04:1 [1].


II. Technical Features of the GLM Architecture: A Differentiated Path from GPT

Zhipu AI’s core technical asset is its independently developed

GLM (General Language Model) architecture
, which is fundamentally different from OpenAI’s GPT technical route [3]:

2.1 Fundamental Innovation in Pre-Training Objectives
Feature
GLM Architecture
GPT Architecture
Pre-Training Objective
Autoregressive Blank Filling Unidirectional Autoregressive Language Modeling
Context Modeling
Bidirectional Context Modeling Pure Forward Generation
Generation and Comprehension
Balances generation capabilities and comprehension performance Focuses on generation capabilities
Chinese Context
Stronger language comprehension ability Relatively weak

The core innovation of GLM lies in: during the pre-training phase, by ‘masking partial text segments and having the model predict them’, the model can learn both generation capabilities (predicting the masked parts) and comprehension capabilities (understanding contextual semantics) [3]. This design gives the model unique advantages in tasks such as long text processing, code generation, and multi-turn dialogue.

2.2 Key Technical Improvements

GLM architecture has made systematic innovations in multiple technical dimensions:

(1) Attention Mechanism Optimization

  • Adopts an improved attention masking strategy to enable more flexible contextual interaction
  • Supports bidirectional attention between Part A (known content) and Part B (content to be predicted)

(2) Rotary Position Embedding (RoPE)

  • Replaces traditional sine encoding with rotary position embedding, significantly enhancing long sequence processing capabilities
  • Achieves excellent
    extrapolability
    — even if the model is trained with limited sequence length, it can effectively process longer inputs [4]
  • Supports remote attenuation characteristics, enhancing the model’s ability to capture long-distance dependencies

(3) Training Strategy Innovation

  • Adopts a multi-stage pre-training strategy
  • Conducts targeted optimizations for different task types
  • Introduces reinforcement learning frameworks (such as the slime framework designed specifically for Agent tasks)
2.3 MoE Architecture Upgrade

The newly released GLM-4.5 series adopts the

Mixture of Experts (MoE) Architecture
:

Parameter Indicators GLM-4.5
Total Parameters Approx. 355B (355 Billion)
Active Parameters 32B (32 Billion)
Context Length 128K
Overall Ranking 3rd among 12 mainstream global benchmarks, 1st among domestic models
Coding Capability Surpasses Qwen3-Coder with an 80.8% win rate
Tool Call Success Rate 90.6%

This architecture achieves a balance of ‘high performance + low resource consumption’, making it particularly suitable for high-concurrency commercial deployment scenarios [5].


III. How the GLM Architecture Supports a Gross Margin of Over 50%

The core logic behind Zhipu AI’s maintenance of a gross margin of over 50% is:

the organic combination of cost advantages brought by technological innovation and business model design
.

3.1 Revenue Structure Analysis

Zhipu’s revenue mainly comes from two segments, with significant differences in their gross margins [2]:

Revenue Segment 2024 Revenue Proportion Gross Margin Proportion in H1 2025
Localized Deployment
CNH 263.9 million 84.5% 59%-68% 84.8%
Cloud Deployment (MaaS)
CNH 48.5 million 15.5% 76.1% → 3.4% →
-0.4%
15.2%

Key Insights
:

  • Localized deployment
    contributes the majority of revenue and gross profit, with a stable gross margin of around 60%, making it a high-margin business
  • Cloud deployment
    saw its gross margin plummet from 76.1% to negative territory due to domestic price wars, with the main goal of seizing market share and developer ecosystem
3.2 Core Mechanisms of the GLM Architecture Supporting High Gross Margins

(1) Generalization Capabilities Reduce Marginal Costs

The “generalization” design of the GLM architecture allows the model to quickly adapt to vertical scenarios

without incurring additional costs for each project
[6]. This is specifically reflected in:

  • A single general model can serve over 12,000 enterprise customers, covering multiple industries such as the internet, public services, telecommunications, consumer electronics, retail, and media
  • Customer repurchase rate exceeds 70%
  • Marginal service costs are extremely low, with significant economies of scale

(2) Domestic Chip Adaptation Reduces Computing Power Costs

The GLM architecture has completed in-depth adaptation to

over 40 domestic chip models
[3], including:

  • Cambricon (FP8+Int4 hybrid quantization deployment)
  • Moore Threads (stable operation with native FP8 precision)

In the current international environment, this capability is strategically significant — it not only ensures supply chain security but also lays the foundation for future cost optimization. Currently, computing power costs account for 71.8% of R&D investment, and domestic chip substitution will be the key to a cost inflection point [2].

(3) Economies of Scale of the MaaS Platform

The MaaS (Model as a Service) platform has

over 2.7 million enterprise and application developers
[2], showing an exponential growth trend:

  • Paid API revenue exceeds the sum of all domestic models
  • The GLM Coding package (priced at 1/7 of Claude’s) acquired 150,000 paid users in 3 months, with ARR exceeding CNH 100 million
  • The platform’s network effect continuously reduces customer acquisition costs

(4) Differentiated Pricing Strategy

In overseas markets, the company adopts a “high cost-performance ratio + strong capabilities” strategy [7]:

  • The GLM Coding package is priced at
    1/7
    of Anthropic Claude’s
  • Trades price advantages for market share and developer ecosystem
  • Overseas revenue already accounts for 11.6% of total revenue and continues to grow

IV. Dynamic Balance Between R&D Expense Ratio and Gross Margin
4.1 Rationality of High R&D Investment

From the perspective of industry development laws, high R&D investment at the current stage is strategically inevitable:

Comparison Dimension Zhipu AI OpenAI Anthropic
R&D Expense Ratio
8.4:1 1.56:1 1.04:1
Commercialization Stage
Early Stage Mid Stage Mid Stage
Strategic Focus
Technological Catch-Up Market Share Technological Leadership

Zhipu’s management clearly stated: “

The top priority is not to make profits, but to promote its technology
” [3]. It is expected that high R&D investment will continue for the next 3-5 years, trading technological leadership for long-term competitive advantages.

4.2 Underlying Logic of Gross Margin Decline

The decline in gross margin from the 2023 peak of 64.6% to 50% in H1 2025 is not a sign of operational deterioration, but rather an

inevitable result of business model transformation
[2]:

  • Initial Stage (2023)
    : Focused on high-end privatized deployment, with high gross margin but limited scale
  • Transformation Stage (2024-2025)
    : Transforming to MaaS platform and standardized products, sacrificing short-term gross profit for scale
  • Mature Stage (expected after 2026)
    : Economies of scale will be released, and gross margin is expected to stabilize and rebound

V. GLM vs GPT: Competitiveness Comparison and Market Validation
5.1 Technical Performance Comparison
Evaluation Dimension GLM-4.7 GPT-5.2 Remarks
Artificial Analysis Index
68 points (1st among domestic models, 6th globally) - 1st among global open-source models
Code Generation (Code Arena)
1st globally Behind Surpassed OpenAI for the first time
Mathematical Reasoning (AIME24)
91.0 points - Significantly outperforms Claude 4 Opus (75.7)
Coding Test (SW E-bench)
64.2% - Close to Claude 4 Sonnet (70.4%)
5.2 International Recognition

In June 2025, OpenAI listed Zhipu as its

“top competitor”
in the report Chinese Progress at the Front, recognizing its “verifiable, responsible, standardized” technical image in sovereign AI competition [2]. This marks that Zhipu’s technical strength has been recognized by top international players.


VI. Conclusions and Outlook

With a financial structure of

7x R&D expense ratio + gross margin of over 50%
, Zhipu AI illustrates the unique business logic of the large model era:

  1. Technological Aspect
    : The differentiated design of the GLM architecture (autoregressive blank filling + bidirectional context modeling + domestic chip adaptation) has built a technological moat
  2. Commercial Aspect
    : “Generalization” capabilities result in extremely low marginal costs, supporting high gross margins; the network effect of the MaaS platform accumulates momentum for long-term growth
  3. Strategic Aspect
    : Sacrificing short-term profits to gain technological leadership and market share, and achieving profitability through economies of scale in the long term

The current R&D expense ratio, which is 7-8 times revenue, is a

necessary investment in the AGI marathon
. With the acceleration of domestic computing power substitution and the release of economies of scale on the MaaS platform, Zhipu is expected to achieve a dynamic balance between R&D investment and commercialization within 3-5 years, truly entering the profit-making stage.


References

[1] Touzijie - “Zhipu AI, Ranked ‘Second’” (https://news.pedaily.cn/202512/558915.shtml)

[2] UniFuncs - “In-Depth Research Report on Zhipu AI: A Panoramic Analysis of Technology, Capital, and Commercialization Paths” (https://unifuncs.com/s/jDsMFzqJ)

[3] Guancha.cn - “The ‘Chinese OpenAI’ from Tsinghua Laboratory” (https://user.guancha.cn/main/content?id=1582269)

[4] Beijing AI Institute - “Understand Rotary Position Embedding (RoPE) in 10 Minutes” (https://hub.baai.ac.cn/view/29979)

[5] Chinaz.com - “2025 Global AI Large Model Recommendation List: In-Depth Comparison of GLM-4.5 vs Qwen3-235B-A22B” (https://www.sohu.com/a/919507513_114774)

[6] The Paper - “‘Chinese Version of OpenAI’ Zhipu Goes Public, the Value of China’s AI Large Models Continues to Rise” (https://m.thepaper.cn/newsDetail_forward_32343111)

[7] Sina Finance - “Beijing’s First Large Model Stock: Zhipu Races for Hong Kong IPO, Raising Over CNH 8 Billion with Annual Revenue Exceeding CNH 300 Million” (https://finance.sina.com.cn/stock/t/2025-12-20/doc-inhcnchz4454223.shtml)

Related Reading Recommendations
No recommended articles
Ask based on this news for deep analysis...
Alpha Deep Research
Auto Accept Plan

Insights are generated using AI models and historical data for informational purposes only. They do not constitute investment advice or recommendations. Past performance is not indicative of future results.