tencent cloud

Tencent Cloud Observability Platform

Release Notes and Announcements
Release Notes
Product Introduction
Overview
Strengths
Basic Features
Basic Concepts
Use Cases
Use Limits
Purchase Guide
Tencent Cloud Product Monitoring
Application Performance Management
Mobile App Performance Monitoring
Real User Monitoring
Cloud Automated Testing
Prometheus Monitoring
Grafana
EventBridge
PTS
Quick Start
Monitoring Overview
Instance Group
Tencent Cloud Product Monitoring
Application Performance Management
Real User Monitoring
Cloud Automated Testing
Performance Testing Service
Prometheus Getting Started
Grafana
Dashboard Creation
EventBridge
Alarm Service
Cloud Product Monitoring
Tencent Cloud Service Metrics
Operation Guide
CVM Agents
Cloud Product Monitoring Integration with Grafana
Troubleshooting
Practical Tutorial
Application Performance Management
Product Introduction
Access Guide
Operation Guide
Practical Tutorial
Parameter Information
FAQs
Mobile App Performance Monitoring
Overview
Operation Guide
Access Guide
Practical Tutorial
Tencent Cloud Real User Monitoring
Product Introduction
Operation Guide
Connection Guide
FAQs
Cloud Automated Testing
Product Introduction
Operation Guide
FAQs
Performance Testing Service
Overview
Operation Guide
Practice Tutorial
JavaScript API List
FAQs
Prometheus Monitoring
Product Introduction
Access Guide
Operation Guide
Practical Tutorial
Terraform
FAQs
Grafana
Product Introduction
Operation Guide
Guide on Grafana Common Features
FAQs
Dashboard
Overview
Operation Guide
Alarm Management
Console Operation Guide
Troubleshooting
FAQs
EventBridge
Product Introduction
Operation Guide
Practical Tutorial
FAQs
Report Management
FAQs
General
Alarm Service
Concepts
Monitoring Charts
CVM Agents
Dynamic Alarm Threshold
CM Connection to Grafana
Documentation Guide
Related Agreements
Application Performance Management Service Level Agreement
APM Privacy Policy
APM Data Processing And Security Agreement
RUM Service Level Agreement
Mobile Performance Monitoring Service Level Agreement
Cloud Automated Testing Service Level Agreement
Prometheus Service Level Agreement
TCMG Service Level Agreements
PTS Service Level Agreement
PTS Use Limits
Cloud Monitor Service Level Agreement
API Documentation
History
Introduction
API Category
Making API Requests
Monitoring Data Query APIs
Alarm APIs
Legacy Alert APIs
Notification Template APIs
TMP APIs
Grafana Service APIs
Event Center APIs
TencentCloud Managed Service for Prometheus APIs
Monitoring APIs
Data Types
Error Codes
Glossary

Accessing LLM Application

PDF
Mode fokus
Ukuran font
Terakhir diperbarui: 2025-10-13 19:10:48
In recent years, LLMs have shown a trend of rapid development. With the rise of DeepSeek, AI application barriers have been further reduced, introducing efficient and low-cost solutions across various industries, promoting an explosion in AI implementation.
The operation of LLM applications involves multiple components and complex interaction processes. Application Performance Management (APM) uses distributed tracing technology to clearly show the call trace of requests between components. In the event of a failure, it enables quick location of the specific issue. Meanwhile, APM can monitor the running state of model applications in real time, detect exceptions timely, and report alarms, allowing O&M personnel to take timely measures to avoid escalation.

Supported LLM Components and Frameworks

Tencent Cloud's self-developed Python probe simultaneously supports automatic event tracking for common LLM frameworks and traditional (non-LLM) Python frameworks, compatible with OpenTelemetry protocol standard, enabling interconnectivity with other applications using OpenTelemetry solution, and supports the following components or frameworks.
Note:
Requires Python 3.9 and above versions.
Supported Components and Frameworks
Link
LLM Components and Frameworks
OpenAI SDK (openai >= 0.27.0): An officially provided API encapsulation by OpenAI for direct calls to all large models compatible with OpenAI standards.
Ollama (ollama >= 0.4.0): A lightweight reasoning framework to run locally and manage open-source large models.
LangChain / LangGraph (langchain-core > 0.1.0): A workflow framework to build and orchestrate large model applications, supporting complex chains and status management.
LlamaIndex (llama-index >= 0.7.0, llama-index-core >= 0.7.0): A RAG framework focused on external data access for LLMs, providing retrieve and index power.
Traditional Python components and frameworks

Access Process

Note:
If the service is deployed in a TKE cluster and the cluster has the Tencent Cloud APM operator installed, manually add the environment variable OTEL_EXPORTER_METRICS_TEMPORALITY_PREFERENCE with the value true in the workload.

Getting Access Point and Token

1. Log in to the TCOP console.
2. In the left menu bar, select APM, then click Application list > Access application.
3. In the Access application drawer frame that pops up on the right, click the Python language.
4. On the Access Python application page, select the Region and Business System.
5. Select OpenTelemetry as Access protocol type.
6. Select a Reporting method through which you want to report data, and obtain your Access Point and Token.
Note:
Report over private network: This reporting method requires your service to run in the Tencent Cloud VPC. The direct connection through VPC can help avoid the security risks of public network communication and save costs on reporting traffic.
Report over public network: If your service is deployed locally or in non-Tencent Cloud VPC, you can report data in this method. However, it involves security risks in public network communication and incurs reporting traffic fees.

Installing pip Packages

Install Tencent Cloud's self-developed probe via pip command, including OpenTelemetry-SDK related dependencies.
pip install tapm-distro opentelemetry-exporter-otlp==1.34.1

tapm-bootstrap -a install

Command Line Method Reporting

Add the tapm-instrument prefix to complete instrumentation and startup. Assuming the original project start command was python app.py, you can now execute the following command to start the Python application.
tapm-instrument --traces_exporter otlp \\
--metrics_exporter otlp \\
--logs_exporter none \\
--service_name <service_name> \\
--resource_attributes "token=<token>,host.name=<host.name>" \\
--exporter_otlp_endpoint <endpoint> \\
python app.py
The corresponding field descriptions are as follows. Replace based on the actual situation.
<service_name>: Application name. Multiple application processes connecting with the same serviceName are displayed as multiple instances under the same application in APM. The application name can contain up to 63 characters, can only contain lowercase letters, digits, and the separator "-", and must start with a lowercase letter and end with a digit or lowercase letter.
<token>: business system Token obtained in preliminary steps.
<host.name>: The hostname of this instance, which is the unique identifier of the application instance. It can usually be set to the IP address of the application instance.
<endpoint>: The access point obtained in the preliminary steps.

Access Verification

After completing the access work, start up the LLM application, and the application will report link data to APM. The connected application will be displayed on the Application Performance Monitoring > LLM Observability > Application list page. Since there is latency in the processing of observable data, if the application or instance does not appear in the console after connecting, please wait about 30 seconds.
Meanwhile, in Application Performance Monitoring > Distributed tracing, you can also query related Span records. By clicking the URL corresponding to the traceID in the first column, you can enter the linkage detail view to analyze each stage of the linkage.

Bantuan dan Dukungan

Apakah halaman ini membantu?

masukan