History
Introduction
API Category
Making API Requests
Region APIs
Instance APIs
Cloud Hosting Cluster APIs
Image APIs
Instance Launch Template APIs
Placement Group APIs
Key APIs
Security Group APIs
Network APIs
Data Types
Error Codes
ci:CreateAsrJobs permission to the action in the authorization policy. For all supported API operations in Cloud Infinite, see CI action.cam:passrole permission. The asynchronous processing interface performs COS read-write operations through cam roles. The passrole permission is used for role passing. For details, refer to Access Management > Write Operation > passrole API.POST /jobs HTTP/1.1Host: <BucketName-APPID>.ci.<Region>.myqcloud.comDate: <GMT Date>Authorization: <Auth String>Content-Length: <length>Content-Type: application/xml<body>
<Request><Tag>SpeechRecognition</Tag><Input><Object>input/test.mp3</Object></Input><Operation><TemplateId>t1460606b9752148c4ab182f55163ba7cd</TemplateId><Output><Region>ap-chongqing</Region><Bucket>test-123456789</Bucket><Object>output/asr.txt</Object></Output><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><CallBack>http://callback.demo.com</CallBack><CallBackFormat>JSON</CallBackFormat></Request>
Node Name (Keyword) | Parent Node | Description | Type | Required or Not |
Request | None. | Container for saving requests | Container | Yes |
Node Name (Keyword) | Parent Node | Description | Type | Required or Not |
Tag | Request | Create task Tag: SpeechRecognition | String | Yes |
Input | Request | Object information to be operated | Container | Yes |
Operation | Request | Operation rule | Container | Yes |
CallBackFormat | Request | Job callback format, JSON or XML, default XML, priority is higher than queue callback format | String | No |
CallBackType | Request | Job callback type, Url or TDMQ, default Url, priority is higher than queue callback type | String | No |
CallBack | Request | Job callback address, priority is higher than queue callback address. When set to no, it means the queue callback address does not generate callbacks. | String | No |
CallBackMqConfig | Request | Task callback TDMQ configuration, required when CallBackType is TDMQ. For details, see CallBackMqConfig | Container | No |
Node Name (Keyword) | Parent Node | Description | Type | Required or Not |
Object | Request.Input | File path | String | No |
Node Name (Keyword) | Parent Node | Description | Type | Required or Not |
TemplateId | Request.Operation | String | No | |
SpeechRecognition | Request.Operation | Container | No | |
Output | Request.Operation | Output Configuration | Container | Yes |
UserData | Request.Operation | Pass through user information, printable ASCII, length not exceeding 1024 | String | No |
JobLevel | Request.Operation | Task priority, level limit: 0, 1, 2. The larger the level, the higher the task priority. Default is 0. | String | No |
Node Name (Keyword) | Parent Node | Description | Type | Required or Not |
Region | Request.Operation.Output | Bucket Region | String | Yes |
Bucket | Request.Operation.Output | Bucket for result storage | String | Yes |
Object | Request.Operation.Output | File name of the result | String | Yes |
<Response><JobsDetail><Code>Success</Code><CreationTime>2021-08-05T15:43:50+0800</CreationTime><EndTime>-</EndTime><Input><BucketId>test-1234567890</BucketId><Object>input/test.mp3</Object><Region>ap-chongqing</Region></Input><JobId>s58ccb634149211ed84ce2b1cd7fbb14a</JobId><Message/><Operation><Output><Bucket>test-1234567890</Bucket><Object>output/asr.txt</Object><Region>ap-chongqing</Region></Output><TemplateId>t1460606b9752148c4ab182f55163ba7cd</TemplateId><TemplateName>speech_demo</TemplateName><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><QueueId>pcd463e1467964d39ad2d3f66aacd8199</QueueId><QueueType>Speeching</QueueType><StartTime>-</StartTime><State>Submitted</State><Tag>SpeechRecognition</Tag></JobsDetail></Response>
Node Name (Keyword) | Parent Node | Description | Type |
Response | None. | Container for saving results | Container |
Node Name (Keyword) | Parent Node | Description | Type |
JobsDetail | Response | Task Details | Container array |
Node Name (Keyword) | Parent Node | Description | Type |
Code | Response.JobsDetail | Error code, only meaningful when State is Failed | String |
Message | Response.JobsDetail | Error description, only meaningful when State is Failed | String |
JobId | Response.JobsDetail | ID of the newly created task | String |
Tag | Response.JobsDetail | Tag of the newly created task: SpeechRecognition | String |
State | Response.JobsDetail | Task Status Submitted: submitted, pending execution Running: executing Success: execution successful Failed: execution failed Pause: task is paused. When the pause queue is triggered, to be executed tasks change to paused state. Cancel: task cancelled | String |
CreationTime | Response.JobsDetail | TaskTask creation time | String |
StartTime | Response.JobsDetail | Task Start Time | String |
EndTime | Response.JobsDetail | Task Closed At | String |
QueueId | Response.JobsDetail | String | |
QueueType | Response.JobsDetail | Queue type of the task | String |
Input | Response.JobsDetail | Input resource address of the task | Container |
Operation | Response.JobsDetail | Rule of the task | Container |
Node Name (Keyword) | Parent Node | Description | Type |
Region | Response.JobsDetail.Input | Bucket Region | String |
BucketId | Response.JobsDetail.Input | The bucket where the source file resides | String |
Object | Response.JobsDetail.Input | Filename of the source file | String |
Node Name (Keyword) | Parent Node | Description | Type |
TemplateId | Response.JobsDetail.Operation | Template ID of the task | String |
TemplateName | Response.JobsDetail.Operation | Template name of the task, return when TemplateId exists | String |
SpeechRecognition | Response.JobsDetail.Operation | Container | |
Output | Response.JobsDetail.Operation | Container | |
UserData | Response.JobsDetail.Operation | Pass through user information | String |
JobLevel | Response.JobsDetail.Operation | Task priority. | String |
SpeechRecognitionResult | Response.JobsDetail.Operation | ASR task result, do not return if none | Container |
Node Name (Keyword) | Parent Node | Description | Type |
AudioTime | Response.JobsDetail.Operation.SpeechRecognitionResult | audio duration (seconds) | String |
Result | Response.JobsDetail.Operation.SpeechRecognitionResult | SpeechRecognitionResult | String |
FlashResult | Response.JobsDetail.Operation.SpeechRecognitionResult | Ultra-fast ASR result | Container array |
ResultDetail | Response.JobsDetail.Operation.SpeechRecognitionResult | Recognition result details, including word time offsets for each sentence, generally used in subtitle generation scenarios. (This field is not null when ResTextFormat=1 in the speech recognition request.) Note: This field may be null, indicating that no valid value can be obtained. | Container array |
Node Name (Keyword) | Parent Node | Description | Type |
channel_id | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult | Sound channel flag, starting from 0, corresponds to the number of audio channels | Int |
text | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult | Sound channel audio integrity recognition result | String |
sentence_list | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult | Sentence/paragraph-level recognition result list | Container array |
Node Name (Keyword) | Parent Node | Description | Type |
text | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list | Sentence/paragraph-level text | String |
start_time | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list | Start Time | Int |
end_time | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list | End Time | Int |
speaker_id | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list | speaker Id (if speaker_diarization is set up in request, distinguish speakers by speaker_Id) | Int |
word_list | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list | Word-level recognition result list | Container array |
Node Name (Keyword) | Parent Node | Description | Type |
word | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list.word_list | word-level text | String |
start_time | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list.word_list | Start Time | Int |
end_time | Response.JobsDetail.Operation.SpeechRecognitionResult.FlashResult.sentence_list.word_list | End Time | Int |
Node Name (Keyword) | Parent Node | Description | Type |
FinalSentence | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Final recognition result of a sentence | String |
SliceSentence | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | intermediate recognition result of a sentence, split into multiple words using space | String |
StartMs | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Start time of a sentence (ms) | String |
EndMs | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | End time of a sentence (ms) | String |
WordsNum | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Number of words in a sentence | String |
SpeechSpeed | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Speaking rate of a sentence, unit: words/sec | String |
SpeakerId | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Sound channel or speaker Id (if speaker_diarization is set up or ChannelNum is set to 2 (stereo), distinguish speakers or channels) | String |
Words | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail | Word details in a sentence | Container array |
Node Name (Keyword) | Parent Node | Description | Type |
Word | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail.Words | word text | String |
OffsetStartMs | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail.Words | Start time offset in a sentence | String |
OffsetEndMs | Response.JobsDetail.Operation.SpeechRecognitionResult.ResultDetail.Words | End time offset in a sentence | String |
POST /jobs HTTP/1.1Authorization: q-sign-algorithm=sha1&q-ak=************************************&q-sign-time=1497530202;1497610202&q-key-time=1497530202;1497610202&q-header-list=&q-url-param-list=&q-signature=****************************************Host: test-1234567890.ci.ap-beijing.myqcloud.comContent-Length: 166Content-Type: application/xml<Request><Tag>SpeechRecognition</Tag><Input><Object>input/test.mp3</Object></Input><Operation><TemplateId>t1460606b9752148c4ab182f55163ba7cd</TemplateId><Output><Region>ap-chongqing</Region><Bucket>test-123456789</Bucket><Object>output/asr.txt</Object></Output><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><CallBack>http://callback.demo.com</CallBack><CallBackFormat>JSON</CallBackFormat></Request>
HTTP/1.1 200 OKContent-Type: application/xmlContent-Length: 230Connection: keep-aliveDate: Mon, 28 Jun 2022 15:23:12 GMTServer: tencent-cix-ci-request-id: NTk0MjdmODlfMjQ4OGY3XzYzYzhf****<Response><JobsDetail><Code>Success</Code><CreationTime>2021-08-05T15:43:50+0800</CreationTime><EndTime>-</EndTime><Input><BucketId>test-1234567890</BucketId><Object>input/test.mp3</Object><Region>ap-chongqing</Region></Input><JobId>s58ccb634149211ed84ce2b1cd7fbb14a</JobId><Message/><Operation><JobLevel>0</JobLevel><Output><Bucket>test-1234567890</Bucket><Object>output/asr.txt</Object><Region>ap-chongqing</Region></Output><TemplateId>t1460606b9752148c4ab182f55163ba7cd</TemplateId><TemplateName>speech_demo</TemplateName><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><QueueId>pcd463e1467964d39ad2d3f66aacd8199</QueueId><QueueType>Speeching</QueueType><StartTime>-</StartTime><State>Submitted</State><Tag>SpeechRecognition</Tag></JobsDetail></Response>
POST /jobs HTTP/1.1Authorization: q-sign-algorithm=sha1&q-ak=************************************&q-sign-time=1497530202;1497610202&q-key-time=1497530202;1497610202&q-header-list=&q-url-param-list=&q-signature=****************************************Host: test-1234567890.ci.ap-beijing.myqcloud.comContent-Length: 166Content-Type: application/xml<Request><Tag>SpeechRecognition</Tag><Input><Object>input/test.mp3</Object></Input><Operation><SpeechRecognition><EngineModelType>16k_zh_video</EngineModelType><ChannelNum>1</ChannelNum><FilterDirty>1</ChannelNum><FilterModal>1</ChannelNum></SpeechRecognition><Output><Region>ap-chongqing</Region><Bucket>test-123456789</Bucket><Object>output/asr.txt</Object></Output><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><CallBack>http://callback.demo.com</CallBack><CallBackFormat>JSON</CallBackFormat></Request>
HTTP/1.1 200 OKContent-Type: application/xmlContent-Length: 230Connection: keep-aliveDate: Mon, 28 Jun 2022 15:23:12 GMTServer: tencent-cix-ci-request-id: NTk0MjdmODlfMjQ4OGY3XzYzYzhf****<Response><JobsDetail><Code>Success</Code><CreationTime>2021-08-05T15:43:50+0800</CreationTime><EndTime>-</EndTime><Input><BucketId>test-1234567890</BucketId><Object>input/test.mp3</Object><Region>ap-chongqing</Region></Input><JobId>s58ccb634149211ed84ce2b1cd7fbb14a</JobId><Message/><Operation><Output><Bucket>test-1234567890</Bucket><Object>output/asr.txt</Object><Region>ap-chongqing</Region></Output><SpeechRecognition><ChannelNum>1</ChannelNum><ConvertNumMode>0</ConvertNumMode><EngineModelType>16k_zh_video</EngineModelType><FilterDirty>0</FilterDirty><FilterModal>0</FilterModal><FilterPunc>0</FilterPunc><OutputFileType>txt</OutputFileType><ResTextFormat>0</ResTextFormat><SpeakerDiarization>0</SpeakerDiarization><SpeakerNumber>0</SpeakerNumber></SpeechRecognition><UserData>This is my data.</UserData><JobLevel>0</JobLevel></Operation><QueueId>pcd463e1467964d39ad2d3f66aacd8199</QueueId><QueueType>Speeching</QueueType><StartTime>-</StartTime><State>Submitted</State><Tag>SpeechRecognition</Tag></JobsDetail></Response>
Apakah halaman ini membantu?
Anda juga dapat Menghubungi Penjualan atau Mengirimkan Tiket untuk meminta bantuan.
masukan