Skip to main content
Use Cases

Industry Cases

Covering finance, energy, transportation, education, healthcare, telecom, government and defense, and more — 12+ flagship deployments trusted by many customers.

12+
Flagship Cases
50+
Industry Customers
7
Key Industries
30%+
Utilization Gain
Case 01 Finance

A Leading Securities Firm — Domestic GPU Slicing + Multi-Cluster Management

Entry
统一访问入口
Control · 管控集群
统一调度
统一监控
统一权限
昇腾测试
TEST
智算平台
容器平台
GPU 节点
昇腾生产
PROD
智算平台
容器平台
GPU 节点
Nvidia 测试
TEST
智算平台
容器平台
GPU 节点
Nvidia 生产
PROD
智算平台
容器平台
GPU 节点
Integration
对接现有系统
LDAP 认证
镜像仓库
对象/NAS 存储
Prometheus
PGSQL
日志系统

Customer Background

The existing container platforms managed only traditional workloads, GPU resources could not be scheduled across clusters, and monitoring, storage, and logging infrastructure was being duplicated.

Customer Pain Points

  • 01 GPUs across multiple clusters cannot be scheduled uniformly
  • 02 Domestic GPUs need finer slicing and scheduling capability
  • 03 Duplicated monitoring, storage, and logging infrastructure

Solution

Adopting a control-plane / GPU-compute separation architecture, using Rise CAMP to integrate multiple domestic and NVIDIA clusters with the existing container platforms, reusing LDAP, image registry, storage, and monitoring, and enabling dynamic slicing on domestic GPUs.

Impact 01
Minutes
Helm-based one-click deployment in minutes
Impact 02
vNPU Slicing
Fine-grained vNPU scheduling on domestic GPUs
Impact 03
Zero Rebuild
Reuses existing storage, monitoring, and logging
Case 02 Finance

Huatai Futures — AI Training & Inference Full Lifecycle Management

Huatai Futures Co., Ltd.

Application · AI 业务
Agent 智能体
知识库
智能投研
风险管理
更多 AI 场景
Model Gateway · 模型网关
Qwen3
DeepSeek R1
多模态模型
向量模型
HithinkGPT
Rise · AI 智算管理平台
AI 容器实例
模型广场
模型微调
弹性伸缩
统一可观测
算力切分/超分
调度策略
镜像管理
多租户管理
资源回收
RUNTIME
Kubernetes
OS
Kylin v10
DB
OceanBase
Infrastructure · 异构算力
Nvidia 节点
国产算力节点
全栈信创 · 训推一体 · 全生命周期

Customer Background

As large models move into core scenarios, traditional architectures struggle to support multi-team concurrent development and fast iteration, and the compute platform must deeply integrate with the domestic-compatible software stack.

Customer Pain Points

  • 01 Multi-team concurrency and fast iteration place greater demands on compute scheduling
  • 02 Deep compatibility with the domestic-compatible ecosystem is required
  • 03 End-to-end training and inference lifecycle management is needed
  • 04 Small-model fine-tuning capability for specific business scenarios is needed

Solution

Building a full-stack domestic-compatible heterogeneous compute base on Rise ModelX, leveraging Rise VAST fine-grained slicing and dynamic scheduling to deliver one-click standardized workspaces, with a built-in model marketplace and inference engine plus support for business-scenario fine-tuning.

Impact 01
Full-Stack Domestic
Adapted to domestic GPUs and the domestic OS and database stack
Impact 02
One-Click Deploy
Mainstream framework workspaces deployed in minutes
Impact 03
Full Lifecycle
One-stop management across training and inference
Case 03 Finance

A Joint-Stock Bank — Heterogeneous AI Compute Base

外挂统一鉴权系统
Active
北京管理集群
LDAP/IAM 认证
统一用户管理
统一调度管理
统一监控告警
Standby
北京管理集群
LDAP/IAM 认证
统一用户管理
统一调度管理
统一监控告警
主集群
北京 · 丰台
AI 容器实例
推理 · 微调
GPU 切分 · 超分
异构 GPU 节点
业务集群
北京 · 昌平
AI 容器实例
推理 · 微调
GPU 切分 · 超分
异构 GPU 节点
灾备集群
珠海机房
AI 容器实例
推理 · 微调
GPU 切分 · 超分
异构 GPU 节点
两地三机房 · 1% 算力 / 1MiB 显存 · 平滑迁移

Customer Background

Financial AI workloads are driving a surge in compute demand, creating an urgent need for an autonomous and controllable compute base. Multi-team parallel workloads on heterogeneous NVIDIA and domestic GPUs have formed compute silos, and whole-card passthrough leaves utilization low.

Customer Pain Points

  • 01 Resources across the two-region three-data-center topology cannot be managed uniformly
  • 02 Heterogeneous GPU scheduling barriers have formed compute silos
  • 03 Whole-card allocation cannot match fine-grained slicing needs
  • 04 Business workloads require zero-downtime smooth migration

Solution

Using Rise CAMP as the heterogeneous compute base to unify management across the three data centers, hide chip-level differences, support fine-grained slicing down to 1% compute and 1MiB VRAM, and deliver zero-downtime smooth migration.

Impact 01
30%+
Overall GPU utilization lifted by 30%+
Impact 02
3 DCs Unified
Unified management and O&M across two regions and three data centers
Impact 03
1% / 1MiB
Ultra-fine slicing for on-demand allocation
Case 04 Finance

A Leading Bank — Heterogeneous Resource Management

用户/权限统一 · 资源监控
Existing · 图灵集群
图灵应用层
容器服务调度
DAG 算子调度
图灵 K8S
New · AI 集群
AI 应用层
容器调度引擎
算子调度引擎
AI K8S
Rise · 适配层
异构资源池化 · 统一纳管 · 弹性伸缩
Infrastructure · 异构算力
CPU 节点
GPU 节点
NPU 节点
既有 Hadoop 集群对接

Customer Background

With multi-team parallel workloads and no GPU virtualization in place, heterogeneous GPU utilization stayed low, and the bank lacked visualized management and unified development standards.

Customer Pain Points

  • 01 Compute demand keeps growing while GPU utilization has significant room to improve
  • 02 Already-allocated resources lack visualized oversight
  • 03 Compatibility between GPU pooling and existing application environments needs to be strengthened
  • 04 Development frameworks and tooling require unified standards

Solution

Using Rise VAST to consolidate existing compute, delivering unified GPU management, overcommit, and elastic scaling, establishing an operations framework, and standardizing development tooling integration.

Impact 01
30%+
GPU utilization lifted by more than 30%
Impact 02
Agile Iteration
Shorter release cycles, faster model iteration
Impact 03
Standardized Delivery
Unified dev tooling for consistent collaboration
Case 05 Finance

A Bank — Unified Scheduling and Elastic Scaling

Frontend · 应用前端
云文档组件
中间件 / DB
AI 基建
AI 素材库
Gateway
统一模型网关调用
AI Models
公文模型
排版图像
数据助手
校对模型
洞察模型
排版生成
模型市场
模型部署
模型管理
模型监控
Rise · 算力池纳管平台 (Kubernetes)
容器管理
任务管理
规格管理
存储管理
镜像管理
租户管理
配额管理
计量计费
GPU 切分/超分
异构纳管
智能告警
操作审计
Infrastructure · AI 基础设施
NPU 节点
GPU 节点
DCU 节点
虚拟机节点

Customer Background

Each department ran its own dedicated hardware as siloed stacks, with no unified authentication or auto-scaling in place, leaving utilization very low.

Customer Pain Points

  • 01 Resources are scattered across departments and need unified management with heterogeneous scheduling
  • 02 A unified permissions and monitoring framework is needed to support standardized operations
  • 03 Model deployment still relies on manual adjustments and needs automated elastic scaling

Solution

Using Rise CAMP on Kubernetes to unify management of diverse domestic heterogeneous resources, providing container/task/tenant/quota/metering, GPU slicing and overcommit, unified scheduling, and full-stack observability.

Impact 01
3-5×
Overall resource utilization lifted 3-5×
Impact 02
Controlled Access
Standardized request, approval, and quota workflow
Impact 03
Auto Scaling
Automatic scale-out under inference traffic spikes
Case 06 Gov & Defense

National Supercomputing Center in Jinan — Solving the Compute Shortage

National Supercomputing Center in Jinan

Control Plane
AI 算力管理平台
资源池化 · 任务调度
Storage
NFS 视频存储
原始视频持久化
Source
考场监控摄像头
海量并发视频流
Data Plane Hub
万兆以太网交换机
视频分流 · 任务路由
Node N01
分析 App
AI 调度组件
GPU · K8s
Node N02
分析 App
AI 调度组件
GPU · K8s
Node N03
分析 App
AI 调度组件
GPU · K8s
Node N04
分析 App
AI 调度组件
GPU · K8s
视频分析网
算力管理网

Customer Background

The supercomputing center needed precise compute scheduling for scenarios like exam monitoring. The original GPU allocation was by physical server, leaving single-host utilization low, significant idle capacity outside exam periods, and heavy reliance on manual O&M.

Customer Pain Points

  • 01 Whole-node allocation cannot fully unlock the value of high-end cards
  • 02 Large amounts of compute sit idle outside exam periods
  • 03 Resource scheduling relies on manual effort, with considerable O&M overhead

Solution

Deploying Rise CAMP scheduling platform to connect exam video streams and NFS storage, pooling GPUs for task-level allocation, and automatically redirecting idle compute to research workloads outside exam periods.

Impact 01
4×+
Compute utilization lifted by more than 4×
Impact 02
One Pool, Many Uses
Idle compute dynamically redirected to research
Impact 03
Automated O&M
Automation replaces manual work, cutting cost and effort
Case 07 Gov & Defense

CEPREI (Saibao Lab) — Heterogeneous AI MaaS Platform

China Electronic Product Reliability and Environmental Testing Research Institute

Application · AI 业务
Agent 智能体
知识库
多模态缺陷检测
办公智能助手
更多 AI 场景
Model Gateway · 统一 API 网关
Qwen3.5
DeepSeek R1
多模态模型
向量模型
工业质量大模型
Rise · AI 智算管理平台
AI 容器实例
模型广场
模型微调
弹性伸缩
统一可观测
算力切分
调度策略
镜像管理
多租户管理
资源回收
Infrastructure · 纯国产异构
超节点互联
平头哥 PPU 节点
单机算力密度
昇腾 Atlas 超节点
高速互联突破瓶颈
安全可控 · 满血版大模型 · 开箱即用

Customer Background

Under a pure-domestic compute environment, full-scale large models must run privately on-premises, and multi-department heterogeneous domestic hardware can easily lead to resource fragmentation.

Customer Pain Points

  • 01 Pure-domestic compute must support full-scale large models
  • 02 Multi-vendor heterogeneous hardware leads to resource fragmentation
  • 03 A unified model service and fine-tuning environment is needed
  • 04 Cross-department unified metering and attribution is needed

Solution

Rise CAMP seamlessly unifies multi-vendor domestic heterogeneous compute into a single pool, supports fine-grained slicing of compute and VRAM, and exposes a unified API gateway that hides backend complexity and provides standard authentication.

Impact 01
Pure Domestic
Domestic heterogeneous compute runs open-source large models efficiently
Impact 02
Super-Node Interconnect
High-speed interconnect breaks single-server bottlenecks
Impact 03
Ready to Use
Text and multimodal models out of the box
Case 08 Transportation

A Leading Transportation SOE — Solving Compute Usability

Application · 应用场景
智能底检
设备运维
车联网
智能列控
MaaS Layer · 公共能力
数据工坊
训练中心
推理服务
模型中心
镜像中心
资源中心
Platform · Rise 统一异构算力管理
异构算力调度底座
异构算力接入
共享与隔离
统一调度引擎
Infrastructure · 算力底座
加速卡
英伟达 · 昇腾
CPU
ARM · x86
存储
网络
Metering · 内部结算
模型计量计费
Token · API Key
智算资源计费
卡时 · 显存 · 任务
MaaS Bridge
对接电信 MaaS

Customer Background

The group's MaaS platform could not elastically provision compute on demand, and multi-team parallel workloads on heterogeneous NVIDIA and domestic GPUs with whole-card passthrough had formed compute silos.

Customer Pain Points

  • 01 The telecom MaaS platform lacks unified compute scheduling
  • 02 Heterogeneous GPU resources urgently need unified management
  • 03 A metering, billing, and resource monitoring framework is still to be built

Solution

Using Rise CAMP to build a unified heterogeneous compute management platform that connects to the telecom MaaS, with a base supporting multi-vendor accelerator integration.

Impact 01
60%+
Heterogeneous coordination lifts overall utilization above 60%
Impact 02
Efficiency 70%↑
Resource allocation efficiency meaningfully improved
Impact 03
Plug and Play
Seamless connection to the telecom MaaS for elastic access
Case 09 Energy

A Leading Energy Enterprise — Unified Scheduling across Multiple MaaS Platforms

Application · AI 应用层
人脸识别
文字识别
数字人
智能助手
MaaS Layer · 多平台并存
资源孤岛 →
千帆 MaaS
文心 · 百度
阿里 MaaS
通义 · 阿里
小模型平台
YOLO · ResNet
开源大模型
DeepSeek 等
Rise · 统一智算管理与调度
→ 一池多用
跨集群统一调度底座
任务管理
调度策略
异构纳管
算力切分
算力池化
计量计费
IaaS Layer
阿里云 · 多集群 K8s
Infrastructure · 异构算力
GPU 节点
GPU 节点
NPU 节点
NPU 节点

Customer Background

Baidu MaaS and Alibaba MaaS are deployed on different clusters with siloed resource pools that cannot be shared; management granularity stops at the physical card level, making a unified compute platform difficult to build.

Customer Pain Points

  • 01 Multi-vendor MaaS resource pools are isolated and hard to share
  • 02 Management granularity stays at the physical card level and adapts poorly to small-model scenarios
  • 03 A lack of dynamic scheduling leaves room to grow overall utilization

Solution

Using Rise VAST to unify both MaaS platforms, applying heterogeneous pooling and virtualization to connect cross-cluster resource scheduling and deliver a fine-grained, elastic, migratable compute base.

Impact 01
30%+
Target compute utilization of 30%+
Impact 02
Cross-Cluster
Cross-cluster task migration simplifies multi-tenant management
Impact 03
Migratable Tasks
Supports cross-cluster task migration and elastic scheduling
Case 10 Transportation

BYD — Autonomous Driving Foundation Model Compute Base

BYD Company Limited

Workload
智驾大模型分布式训练
Rise · AI 智算管理平台 (K8s + Volcano)
存储管理
镜像管理
AI 容器实例
分布式训练
统一可观测
算力池化
调度策略
超时/自动重试
多租户管理
资源回收
Compute
GPU 训练集群
无阻塞 IB 算力网络
Storage
全闪
高并发热数据
混闪
温冷数据
IB 存储网络
协同调度 · 自动容错 · 冷热分层 · 训练不中断

Customer Background

Large-model training nodes are tightly coupled, and traditional scheduling easily leads to global stalls and idle compute; long-duration training is prone to failures, and the trade-off between heavy road-test data I/O and storage TCO is sharp.

Customer Pain Points

  • 01 Distributed scheduling deadlocks leave compute idle
  • 02 Long-duration training demands high fault tolerance; failures can abort the entire job
  • 03 Tension between massive-data I/O throughput and storage TCO

Solution

Using Rise CAMP on the GPU cluster with the Volcano scheduling framework and automatic fault tolerance, backed by all-flash and hybrid-flash tiered storage for hot and cold data, improving training efficiency and cost control.

Impact 01
Utilization Up
Coordinated scheduling reduces compute fragmentation and interruption
Impact 02
Faster Iteration
Stable training environment shortens delivery cycles
Impact 03
TCO Optimized
Tiered storage controls investment in road-test data
Case 11 Education

An Industry-Focused University — Agent Platform & Heterogeneous Compute Management

Agent Platform · 智能体平台
科研项目
课程编写
校园规章
数字人
训练实验
安全合规
学生编程
敏感词检测
语音提取
视频处理
Model Gateway · 模型接口 + API Key
Qwen 大模型
向量模型
语音模型
视频模型
编程模型
Shared Pool · 共享池
多模型混部
动态比例切分 · 1/4 · 1/2 任意组合
Dedicated Pool · 独享池
整卡直通
大模型训练 · 高性能任务
Rise · 智算管理平台 (Kubernetes)
任务调度
动态切分
租户隔离
镜像管理
资源监控
Infrastructure · 异构节点
Nvidia 节点
共享 + 独享
昇腾节点
动态切分
昇腾节点
动态切分

Customer Background

Building an aviation-industry large-model incubation base required unified management of multi-vendor GPUs and a visual agent platform integrating various open-source models.

Customer Pain Points

  • 01 NVIDIA and domestic Ascend compute differ significantly at the underlying layer, making heterogeneous cards hard to manage uniformly
  • 02 Small-model development monopolizes whole cards, leaving compute underutilized; finer allocation is needed
  • 03 The existing platform struggled to reliably support high-concurrency access across the campus

Solution

Deploying Rise CAMP and the agent platform, using domestic GPUs for dynamic proportional slicing, and applying a shared/dedicated dual-pool model to unify open-source model management across research, teaching, and digital human scenarios.

Impact 01
30%+
Flexible slicing lifts utilization by 30%+
Impact 02
1000+ Concurrency
Stable 1000+ concurrent Q&A during the new semester
Impact 03
Domestic + Heterogeneous
Unified scheduling across domestic and multi-vendor chips
Case 12 Healthcare

Chongqing 2nd Hospital — Business Agility Boost

The Second Affiliated Hospital of Chongqing Medical University

Role
院区管理员
Role
院区开发人员
Role
各业务厂商
Standard Mgmt
集群管理
显卡分配
规格创建
公共镜像
公共存储
Dev & Test
容器实例
推理服务
分布式训练
开发工具
VSCode · Jupyter · SSH · Web Console
Deploy
模型上传
模型部署
智算服务
GPU 规格
资源监控
Rise · 算力管理平台
标签管理
授权管理
规格管理
计量计费
租户管理
GPU 池化
节点管理
算力超分
Infrastructure
GPU 集群 · 1% 算力 / 1MiB 显存切分

Customer Background

Clinical department data must be strictly isolated, multi-vendor workloads sharing GPUs make whole-card allocation impractical, and lengthy AI environment setup slows delivery.

Customer Pain Points

  • 01 Whole-card allocation does not match the fine-grained compute needs of small models
  • 02 Strong data isolation requirements across departments
  • 03 Whole-card allocation cannot serve multiple vendors sharing the same infrastructure
  • 04 Lengthy AI environment setup impacts development delivery pace

Solution

Deploying Rise CAMP on the GPU cluster to provide GPU pooling with fine-grained slicing and multi-tenant isolation, and integrating development tooling that accelerates AI delivery.

Impact 01
1% / 1MiB
GPU sliced by 1% compute and 1MiB VRAM
Impact 02
Data Isolation
Custom mounts ensure compliance and security
Impact 03
Agile Delivery
Integrated toolchain shortens delivery cycles
More Customers

Beyond these flagship use cases, our heterogeneous AI compute platform deeply empowers over 100 leading organizations across central SOEs, finance, universities, and research institutes. We help customers maximize hardware utilization, lower compute TCO, and accelerate their AI transformation.

Get In Touch

Get a Tailored AI Compute Solution for Your Industry

Contact Us