Infinigence AI
Chinese GPU cloud and LLM inference startup (无问芯穹) offering heterogeneous compute for AI training/inference.
Our Verdict
Interesting bet if you need scale GPU compute inside China's constrained supply chain.
Pros
- Aggregates heterogeneous Chinese GPU supply
- Addresses chip-shortage bottlenecks locally
- Strong research pedigree from Tsinghua
Cons
- Performance varies by underlying hardware pool
- Limited visibility outside Chinese market
- Early-stage — tooling and SLAs still maturing
Best for: China-based AI teams needing GPU capacity when H100s are blocked
Not for: Global teams with access to NVIDIA-standard hyperscaler clouds
When to Use Infinigence AI
Good fit if you need
- Running low-cost GPU inference for LLMs at Chinese datacenters
- Deploying open-source models on affordable Chinese cloud GPU
- Scaling LLM throughput for high-volume consumer AI products
- Accessing Nvidia H800 capacity for ML training in China
Lock-in Assessment
Medium 3/5
Lock-in Score 3/5
Pricing
Price wrong?Infinigence AI Pricing
- Pricing Model
- usage
- Free Tier
- No
- Entry Price
- —
- Enterprise Available
- No
- Transparency Score
- —
Beta — estimates may differ from actual pricing
1,000
1001K10K100K1M
10,000
1K10K100K1M10M
Estimated Monthly Cost
$25
Estimated Annual Cost
$300
Estimates are approximate and may not reflect current pricing. Always check the official pricing page.
Community Discussion
Comments powered by Giscus (GitHub Discussions). You need a GitHub account to comment.