1
1
/*
2
- * (C) Copyright IBM Corp. 2018, 2021 .
2
+ * (C) Copyright IBM Corp. 2022 .
3
3
*
4
4
* Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
5
5
* the License. You may obtain a copy of the License at
@@ -25,11 +25,9 @@ public class CreateJobOptions extends GenericModel {
25
25
26
26
/**
27
27
* The identifier of the model that is to be used for the recognition request. (**Note:** The
28
- * model `ar-AR_BroadbandModel` is deprecated; use `ar-MS_BroadbandModel` instead.) See
29
- * [Previous-generation languages and
30
- * models](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models) and
31
- * [Next-generation languages and
32
- * models](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models-ng).
28
+ * model `ar-AR_BroadbandModel` is deprecated; use `ar-MS_BroadbandModel` instead.) See [Using a
29
+ * model for speech
30
+ * recognition](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models-use).
33
31
*/
34
32
public interface Model {
35
33
/** ar-AR_BroadbandModel. */
@@ -38,6 +36,8 @@ public interface Model {
38
36
String AR_MS_BROADBANDMODEL = "ar-MS_BroadbandModel" ;
39
37
/** ar-MS_Telephony. */
40
38
String AR_MS_TELEPHONY = "ar-MS_Telephony" ;
39
+ /** cs-CZ_Telephony. */
40
+ String CS_CZ_TELEPHONY = "cs-CZ_Telephony" ;
41
41
/** de-DE_BroadbandModel. */
42
42
String DE_DE_BROADBANDMODEL = "de-DE_BroadbandModel" ;
43
43
/** de-DE_NarrowbandModel. */
@@ -46,6 +46,8 @@ public interface Model {
46
46
String DE_DE_TELEPHONY = "de-DE_Telephony" ;
47
47
/** en-AU_BroadbandModel. */
48
48
String EN_AU_BROADBANDMODEL = "en-AU_BroadbandModel" ;
49
+ /** en-AU_Multimedia. */
50
+ String EN_AU_MULTIMEDIA = "en-AU_Multimedia" ;
49
51
/** en-AU_NarrowbandModel. */
50
52
String EN_AU_NARROWBANDMODEL = "en-AU_NarrowbandModel" ;
51
53
/** en-AU_Telephony. */
@@ -54,6 +56,8 @@ public interface Model {
54
56
String EN_IN_TELEPHONY = "en-IN_Telephony" ;
55
57
/** en-GB_BroadbandModel. */
56
58
String EN_GB_BROADBANDMODEL = "en-GB_BroadbandModel" ;
59
+ /** en-GB_Multimedia. */
60
+ String EN_GB_MULTIMEDIA = "en-GB_Multimedia" ;
57
61
/** en-GB_NarrowbandModel. */
58
62
String EN_GB_NARROWBANDMODEL = "en-GB_NarrowbandModel" ;
59
63
/** en-GB_Telephony. */
@@ -68,6 +72,8 @@ public interface Model {
68
72
String EN_US_SHORTFORM_NARROWBANDMODEL = "en-US_ShortForm_NarrowbandModel" ;
69
73
/** en-US_Telephony. */
70
74
String EN_US_TELEPHONY = "en-US_Telephony" ;
75
+ /** en-WW_Medical_Telephony. */
76
+ String EN_WW_MEDICAL_TELEPHONY = "en-WW_Medical_Telephony" ;
71
77
/** es-AR_BroadbandModel. */
72
78
String ES_AR_BROADBANDMODEL = "es-AR_BroadbandModel" ;
73
79
/** es-AR_NarrowbandModel. */
@@ -84,8 +90,12 @@ public interface Model {
84
90
String ES_ES_BROADBANDMODEL = "es-ES_BroadbandModel" ;
85
91
/** es-ES_NarrowbandModel. */
86
92
String ES_ES_NARROWBANDMODEL = "es-ES_NarrowbandModel" ;
93
+ /** es-ES_Multimedia. */
94
+ String ES_ES_MULTIMEDIA = "es-ES_Multimedia" ;
87
95
/** es-ES_Telephony. */
88
96
String ES_ES_TELEPHONY = "es-ES_Telephony" ;
97
+ /** es-LA_Telephony. */
98
+ String ES_LA_TELEPHONY = "es-LA_Telephony" ;
89
99
/** es-MX_BroadbandModel. */
90
100
String ES_MX_BROADBANDMODEL = "es-MX_BroadbandModel" ;
91
101
/** es-MX_NarrowbandModel. */
@@ -136,6 +146,8 @@ public interface Model {
136
146
String NL_NL_BROADBANDMODEL = "nl-NL_BroadbandModel" ;
137
147
/** nl-NL_NarrowbandModel. */
138
148
String NL_NL_NARROWBANDMODEL = "nl-NL_NarrowbandModel" ;
149
+ /** nl-NL_Telephony. */
150
+ String NL_NL_TELEPHONY = "nl-NL_Telephony" ;
139
151
/** pt-BR_BroadbandModel. */
140
152
String PT_BR_BROADBANDMODEL = "pt-BR_BroadbandModel" ;
141
153
/** pt-BR_NarrowbandModel. */
@@ -146,6 +158,8 @@ public interface Model {
146
158
String ZH_CN_BROADBANDMODEL = "zh-CN_BroadbandModel" ;
147
159
/** zh-CN_NarrowbandModel. */
148
160
String ZH_CN_NARROWBANDMODEL = "zh-CN_NarrowbandModel" ;
161
+ /** zh-CN_Telephony. */
162
+ String ZH_CN_TELEPHONY = "zh-CN_Telephony" ;
149
163
}
150
164
151
165
/**
@@ -752,11 +766,9 @@ public String contentType() {
752
766
* Gets the model.
753
767
*
754
768
* <p>The identifier of the model that is to be used for the recognition request. (**Note:** The
755
- * model `ar-AR_BroadbandModel` is deprecated; use `ar-MS_BroadbandModel` instead.) See
756
- * [Previous-generation languages and
757
- * models](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models) and
758
- * [Next-generation languages and
759
- * models](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models-ng).
769
+ * model `ar-AR_BroadbandModel` is deprecated; use `ar-MS_BroadbandModel` instead.) See [Using a
770
+ * model for speech
771
+ * recognition](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-models-use).
760
772
*
761
773
* @return the model
762
774
*/
@@ -1020,7 +1032,9 @@ public Boolean timestamps() {
1020
1032
*
1021
1033
* <p>If `true`, the service filters profanity from all output except for keyword results by
1022
1034
* replacing inappropriate words with a series of asterisks. Set the parameter to `false` to
1023
- * return results with no censoring. Applies to US English and Japanese transcription only. See
1035
+ * return results with no censoring.
1036
+ *
1037
+ * <p>**Note:** The parameter can be used with US English and Japanese transcription only. See
1024
1038
* [Profanity
1025
1039
* filtering](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-formatting#profanity-filtering).
1026
1040
*
@@ -1038,7 +1052,7 @@ public Boolean profanityFilter() {
1038
1052
* final transcript of a recognition request. For US English, the service also converts certain
1039
1053
* keyword strings to punctuation symbols. By default, the service performs no smart formatting.
1040
1054
*
1041
- * <p>**Beta :** The parameter is beta functionality. Applies to US English, Japanese, and Spanish
1055
+ * <p>**Note :** The parameter can be used with US English, Japanese, and Spanish (all dialects)
1042
1056
* transcription only.
1043
1057
*
1044
1058
* <p>See [Smart
@@ -1056,16 +1070,13 @@ public Boolean smartFormatting() {
1056
1070
* <p>If `true`, the response includes labels that identify which words were spoken by which
1057
1071
* participants in a multi-person exchange. By default, the service returns no speaker labels.
1058
1072
* Setting `speaker_labels` to `true` forces the `timestamps` parameter to be `true`, regardless
1059
- * of whether you specify `false` for the parameter.
1060
- *
1061
- * <p>**Beta:** The parameter is beta functionality. * For previous-generation models, the
1062
- * parameter can be used for Australian English, US English, German, Japanese, Korean, and Spanish
1063
- * (both broadband and narrowband models) and UK English (narrowband model) transcription only. *
1064
- * For next-generation models, the parameter can be used for English (Australian, Indian, UK, and
1065
- * US), German, Japanese, Korean, and Spanish transcription only.
1073
+ * of whether you specify `false` for the parameter. * _For previous-generation models,_ the
1074
+ * parameter can be used with Australian English, US English, German, Japanese, Korean, and
1075
+ * Spanish (both broadband and narrowband models) and UK English (narrowband model) transcription
1076
+ * only. * _For next-generation models,_ the parameter can be used with Czech, English
1077
+ * (Australian, Indian, UK, and US), German, Japanese, Korean, and Spanish transcription only.
1066
1078
*
1067
- * <p>Restrictions and limitations apply to the use of speaker labels for both types of models.
1068
- * See [Speaker
1079
+ * <p>See [Speaker
1069
1080
* labels](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-speaker-labels).
1070
1081
*
1071
1082
* @return the speakerLabels
@@ -1096,8 +1107,6 @@ public String customizationId() {
1096
1107
* that are recognized by the specified grammar; it does not recognize other custom words from the
1097
1108
* model's words resource.
1098
1109
*
1099
- * <p>**Beta:** The parameter is beta functionality.
1100
- *
1101
1110
* <p>See [Using a grammar for speech
1102
1111
* recognition](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-grammarUse).
1103
1112
*
@@ -1120,8 +1129,8 @@ public String grammarName() {
1120
1129
* disables keyword spotting (ignores the `keywords` and `keywords_threshold` parameters) and
1121
1130
* returns only a single final transcript (forces the `max_alternatives` parameter to be `1`).
1122
1131
*
1123
- * <p>**Beta :** The parameter is beta functionality. Applies to US English, Japanese, and Korean
1124
- * transcription only.
1132
+ * <p>**Note :** The parameter can be used with US English, Japanese, and Korean transcription
1133
+ * only.
1125
1134
*
1126
1135
* <p>See [Numeric
1127
1136
* redaction](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-formatting#numeric-redaction).
@@ -1222,8 +1231,11 @@ public Double endOfPhraseSilenceTime() {
1222
1231
* semantic features of the input, for example, at the conclusion of meaningful phrases such as
1223
1232
* sentences. The service bases its understanding of semantic features on the base language model
1224
1233
* that you use with a request. Custom language models and grammars can also influence how and
1225
- * where the service splits a transcript. By default, the service splits transcripts based solely
1226
- * on the pause interval.
1234
+ * where the service splits a transcript.
1235
+ *
1236
+ * <p>By default, the service splits transcripts based solely on the pause interval. If the
1237
+ * parameters are used together on the same request, `end_of_phrase_silence_time` has precedence
1238
+ * over `split_transcript_at_phrase_end`.
1227
1239
*
1228
1240
* <p>See [Split transcript at phrase
1229
1241
* end](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-parsing#split-transcript).
@@ -1246,8 +1258,13 @@ public Boolean splitTranscriptAtPhraseEnd() {
1246
1258
* * 0.5 (the default) provides a reasonable compromise for the level of sensitivity. * 1.0
1247
1259
* suppresses no audio (speech detection sensitivity is disabled).
1248
1260
*
1249
- * <p>The values increase on a monotonic curve. See [Speech detector
1250
- * sensitivity](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-parameters-sensitivity).
1261
+ * <p>The values increase on a monotonic curve.
1262
+ *
1263
+ * <p>The parameter is supported with all next-generation models and with most previous-generation
1264
+ * models. See [Speech detector
1265
+ * sensitivity](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-parameters-sensitivity)
1266
+ * and [Language model
1267
+ * support](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-support).
1251
1268
*
1252
1269
* @return the speechDetectorSensitivity
1253
1270
*/
@@ -1266,8 +1283,13 @@ public Float speechDetectorSensitivity() {
1266
1283
* (background audio suppression is disabled). * 0.5 provides a reasonable level of audio
1267
1284
* suppression for general usage. * 1.0 suppresses all audio (no audio is transcribed).
1268
1285
*
1269
- * <p>The values increase on a monotonic curve. See [Background audio
1270
- * suppression](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-parameters-suppression).
1286
+ * <p>The values increase on a monotonic curve.
1287
+ *
1288
+ * <p>The parameter is supported with all next-generation models and with most previous-generation
1289
+ * models. See [Background audio
1290
+ * suppression](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-parameters-suppression)
1291
+ * and [Language model
1292
+ * support](https://cloud.ibm.com/docs/speech-to-text?topic=speech-to-text-detection#detection-support).
1271
1293
*
1272
1294
* @return the backgroundAudioSuppression
1273
1295
*/
0 commit comments