tencent cloud

Tencent Cloud EdgeOne

Edge inference fee (postpaid)

PDF
Focus Mode
Font Size
Last updated: 2026-04-15 17:34:30
Edge Inference provides GPU inference services based on EdgeOne edge nodes, allowing users to deploy custom model images or platform-preset models to edge nodes for inference. Edge Inference is only supported in the Enterprise Edition plan and adopts a postpaid billing method based on instance usage duration. With a single inference instance as the smallest billing unit, postpaid bills are generated based on the instance running duration.
Instance running duration: refers to the total time (in seconds) from the startup to the termination of an inference instance. Billing is calculated per second with a minimum charge of 1 second, and any fraction of a second will be rounded up to 1 second.
Note:
The Edge Inference feature is currently in beta testing and requires being added to the allowlist for access. It is only supported in the Enterprise Edition plan. If needed, please Contact Us.

Edge Inference Cost

Edge Inference is billed on a postpaid basis according to the running duration of instances with different GPU specifications and settled monthly. The billing method and pricing for postpaid are as follows:
Billable Item
GPU specifications
List price (USD/second)
Billing Modes
Settlement Cycle
Custom Inference Service
Entry-level (A-tier)
0.000217 USD/second
Postpaid
Month
Basic (B-tier)
0.000220 USD/second
Basic Enhanced (C-tier)
0.000250 USD/second
Note:
1. Edge Inference is billed based on the actual running duration of instances. Billing stops as soon as the instance is stopped.
2. If AS is enabled, each instance is independently metered, and the total cost is the cumulative value of the running duration costs for all instances.


Help and Support

Was this page helpful?

Help us improve! Rate your documentation experience in 5 mins.

Feedback