This API is used to create a speech recognition template.
POST /template HTTP/1.1
Host: <BucketName-APPID>.ci.<Region>.myqcloud.com
Date: <GMT Date>
Authorization: <Auth String>
Content-Length: <length>
Content-Type: application/xml
<body>
Note:
- Authorization: Auth String (for more information, see Request Signature).
- When this feature is used by a sub-account, relevant permissions must be granted.
This API only uses common request headers. For more information, see Common Request Headers.
This request requires the following request body:
<Request>
<Tag>SpeechRecognition</Tag>
<Name>TemplateName</Name>
<SpeechRecognition>
<EngineModelType>16k_zh</EngineModelType>
<ResTextFormat>1</ResTextFormat>
<FilterDirty>0</FilterDirty>
<FilterModal>1</FilterModal>
<ConvertNumMode>0</ConvertNumMode>
<SpeakerDiarization>1</SpeakerDiarization>
<SpeakerNumber>0</SpeakerNumber>
<FilterPunc>0</FilterPunc>
<OutputFileType>txt</OutputFileType>
</SpeechRecognition>
</Request>
The nodes are described as follows:
Node Name (Keyword) | Parent Node | Description | Type | Required |
---|---|---|---|---|
Request | None | Request container. | Container | Yes |
Request
has the following sub-nodes:
Node Name (Keyword) | Parent Node | Description | Type | Required | Constraints |
---|---|---|---|---|---|
Tag | Request | Template tag: SpeechRecognition. | String | Yes | No |
Name | Request | Template name, which can contain letters, digits, underscores (_), hyphens (-), and asterisks (*). | String | Yes | None |
SpeechRecognition | Request | Speech recognition parameter. | Container | Yes | None |
SpeechRecognition
has the following sub-nodes:
Node Name (Keyword) | Parent Node | Description | Type | Required |
---|---|---|---|---|
EngineModelType | Request.Speech Recognition |
Engine model type, divided into phone call and non-phone call scenarios. Phone call scenarios:
|
String | Yes |
ChannelNum | Request.Speech Recognition |
Number of sound channels:
|
Integer | Yes |
ResTextFormat | Request.Speech Recognition |
Format of the returned recognition result.
|
Integer | Yes |
FilterDirty | Request.Speech Recognition |
Whether to filter restricted words (for the Mandarin engine only).
|
Integer | No |
FilterModal | Request.Speech Recognition |
Whether to filter modal particles (for the Mandarin engine only).
|
Integer | No |
ConvertNumMode | Request.Speech Recognition |
Whether to intelligently convert Chinese numbers to Arabic numerals (for the Mandarin engine only):
|
Integer | No |
SpeakerDiarization | Request.Speech Recognition |
Whether to enable speaker separation:
|
Integer | No |
SpeakerNumber | Request.Speech Recognition |
Number of speakers to be separated (with speaker separation enabled). Value range: 0–10. 0 . |
Integer | No |
FilterPunc | Request.Speech Recognition |
Whether to filter punctuation marks (currently for the Mandarin engine only):
|
Integer | No |
OutputFileType | Request.Speech Recognition |
Output file type. Valid values: txt , srt . Default value: txt . |
String | No |
This API only returns common response headers. For more information, see Common Response Headers.
The response body returns application/xml data. The following contains all the nodes:
<Response>
<Template>
<Tag>SpeechRecognition</Tag>
<Name>TemplateName</Name>
<State>Normal</State>
<Tag>SpeechRecognition</Tag>
<CreateTime></CreateTime>
<UpdateTime></UpdateTime>
<BucketId></BucketId>
<Category>Custom</Category>
<SpeechRecognition>
<EngineModelType>16k_zh</EngineModelType>
<ResTextFormat>1</ResTextFormat>
<FilterDirty>0</FilterDirty>
<FilterModal>1</FilterModal>
<ConvertNumMode>0</ConvertNumMode>
<SpeakerDiarization>1</SpeakerDiarization>
<SpeakerNumber>0</SpeakerNumber>
<FilterPunc>0</FilterPunc>
<OutputFileType>txt</OutputFileType>
</SpeechRecognition>
</Template>
</Response>
The nodes are as described below:
Node Name (Keyword) | Parent Node | Description | Type |
---|---|---|---|
Response | None | Response container | Container |
Response
has the following sub-nodes:
Node Name (Keyword) | Parent Node | Description | Type |
---|---|---|---|
TemplateId | Response.Template | Template ID. | String |
Name | Response.Template | Template name. | String |
BucketId | Response.Template | Template bucket. | String |
Category | Response.Template | Template category: Custom or Official . |
String |
Tag | Response.Template | Template tag: SpeechRecognition. | String |
UpdateTime | Response.Template | Update time. | String |
CreateTime | Response.Template | Creation time. | String |
SpeechRecognition | Response.Template | Same as the Request.SpeechRecognition in the request body. |
Container |
There are no special error messages for this request. For common error messages, see Error Codes.
POST /template HTTP/1.1
Authorization: q-sign-algorithm=sha1&q-ak=AKIDZfbOAo7cllgPvF9cXFrJD0a1ICvR****&q-sign-time=1497530202;1497610202&q-key-time=1497530202;1497610202&q-header-list=&q-url-param-list=&q-signature=28e9a4986df11bed0255e97ff90500557e0e****
Host: test-1234567890.ci.ap-chongqing.myqcloud.com
Content-Length: 1666
Content-Type: application/xml
<Request>
<Tag>SpeechRecognition</Tag>
<Name>TemplateName</Name>
<SpeechRecognition>
<EngineModelType>16k_zh</EngineModelType>
<ResTextFormat>1</ResTextFormat>
<FilterDirty>0</FilterDirty>
<FilterModal>1</FilterModal>
<ConvertNumMode>0</ConvertNumMode>
<SpeakerDiarization>1</SpeakerDiarization>
<SpeakerNumber>0</SpeakerNumber>
<FilterPunc>0</FilterPunc>
<OutputFileType>txt</OutputFileType>
</SpeechRecognition>
</Request>
HTTP/1.1 200 OK
Content-Type: application/xml
Content-Length: 100
Connection: keep-alive
Date: Thu, 14 Jul 2022 12:37:29 GMT
Server: tencent-ci
x-ci-request-id: NTk0MjdmODlfMjQ4OGY3XzYzYzhf****
<Response>
<Template>
<TemplateId>t1460606b9752148c4ab182f55163ba7cd</TemplateId>
<Name>TemplateName</Name>
<State>Normal</State>
<Tag>SpeechRecognition</Tag>
<CreateTime>2020-08-05T11:35:24+0800</CreateTime>
<UpdateTime>2020-08-31T16:15:20+0800</UpdateTime>
<BucketId>test-1234567890</BucketId>
<Category>Custom</Category>
<SpeechRecognition>
<EngineModelType>16k_zh</EngineModelType>
<ChannelNum>1</ChannelNum>
<ResTextFormat>0</ResTextFormat>
<FilterDirty>1</FilterDirty>
<FilterModal>0</FilterModal>
<ConvertNumMode>1</ConvertNumMode>
<SpeakerDiarization>0</SpeakerDiarization>
<SpeakerNumber>0</SpeakerNumber>
<FilterPunc>0</FilterPunc>
</SpeechRecognition>
</Template>
</Response>
Apakah halaman ini membantu?