UNPKG

microsoft-cognitiveservices-speech-sdk

Version:
462 lines (461 loc) 21.4 kB
/** * Defines speech property ids. * @class PropertyId */ export declare enum PropertyId { /** * The Cognitive Services Speech Service subscription Key. If you are using an intent recognizer, you need to * specify the LUIS endpoint key for your particular LUIS app. Under normal circumstances, you shouldn't * have to use this property directly. * Instead, use [[SpeechConfig.fromSubscription]]. * @member PropertyId.SpeechServiceConnection_Key */ SpeechServiceConnection_Key = 0, /** * The Cognitive Services Speech Service endpoint (url). Under normal circumstances, you shouldn't * have to use this property directly. * Instead, use [[SpeechConfig.fromEndpoint]]. * NOTE: This endpoint is not the same as the endpoint used to obtain an access token. * @member PropertyId.SpeechServiceConnection_Endpoint */ SpeechServiceConnection_Endpoint = 1, /** * The Cognitive Services Speech Service region. Under normal circumstances, you shouldn't have to * use this property directly. * Instead, use [[SpeechConfig.fromSubscription]], [[SpeechConfig.fromEndpoint]], [[SpeechConfig.fromAuthorizationToken]]. * @member PropertyId.SpeechServiceConnection_Region */ SpeechServiceConnection_Region = 2, /** * The Cognitive Services Speech Service authorization token (aka access token). Under normal circumstances, * you shouldn't have to use this property directly. * Instead, use [[SpeechConfig.fromAuthorizationToken]], [[SpeechRecognizer.authorizationToken]], * [[IntentRecognizer.authorizationToken]], [[TranslationRecognizer.authorizationToken]], [[SpeakerRecognizer.authorizationToken]]. * @member PropertyId.SpeechServiceAuthorization_Token */ SpeechServiceAuthorization_Token = 3, /** * The Cognitive Services Speech Service authorization type. Currently unused. * @member PropertyId.SpeechServiceAuthorization_Type */ SpeechServiceAuthorization_Type = 4, /** * The Cognitive Services Speech Service endpoint id. Under normal circumstances, you shouldn't * have to use this property directly. * Instead, use [[SpeechConfig.endpointId]]. * NOTE: The endpoint id is available in the Speech Portal, listed under Endpoint Details. * @member PropertyId.SpeechServiceConnection_EndpointId */ SpeechServiceConnection_EndpointId = 5, /** * The list of comma separated languages (BCP-47 format) used as target translation languages. Under normal circumstances, * you shouldn't have to use this property directly. * Instead use [[SpeechTranslationConfig.addTargetLanguage]], * [[SpeechTranslationConfig.targetLanguages]], [[TranslationRecognizer.targetLanguages]]. * @member PropertyId.SpeechServiceConnection_TranslationToLanguages */ SpeechServiceConnection_TranslationToLanguages = 6, /** * The name of the Cognitive Service Text to Speech Service Voice. Under normal circumstances, you shouldn't have to use this * property directly. * Instead, use [[SpeechTranslationConfig.voiceName]]. * NOTE: Valid voice names can be found <a href="https://aka.ms/csspeech/voicenames">here</a>. * @member PropertyId.SpeechServiceConnection_TranslationVoice */ SpeechServiceConnection_TranslationVoice = 7, /** * Translation features. * @member PropertyId.SpeechServiceConnection_TranslationFeatures */ SpeechServiceConnection_TranslationFeatures = 8, /** * The Language Understanding Service Region. Under normal circumstances, you shouldn't have to use this property directly. * Instead, use [[LanguageUnderstandingModel]]. * @member PropertyId.SpeechServiceConnection_IntentRegion */ SpeechServiceConnection_IntentRegion = 9, /** * The host name of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments. * You shouldn't have to use this property directly. * Instead use <see cref="SpeechConfig.SetProxy(string,int,string,string)"/>. * Added in version 1.4.0. */ SpeechServiceConnection_ProxyHostName = 10, /** * The port of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments. * You shouldn't have to use this property directly. * Instead use <see cref="SpeechConfig.SetProxy(string,int,string,string)"/>. * Added in version 1.4.0. */ SpeechServiceConnection_ProxyPort = 11, /** * The user name of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments. * You shouldn't have to use this property directly. * Instead use <see cref="SpeechConfig.SetProxy(string,int,string,string)"/>. * Added in version 1.4.0. */ SpeechServiceConnection_ProxyUserName = 12, /** * The password of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments. * You shouldn't have to use this property directly. * Instead use <see cref="SpeechConfig.SetProxy(string,int,string,string)"/>. * Added in version 1.4.0. */ SpeechServiceConnection_ProxyPassword = 13, /** * The Cognitive Services Speech Service recognition Mode. Can be "INTERACTIVE", "CONVERSATION", "DICTATION". * This property is intended to be read-only. The SDK is using it internally. * @member PropertyId.SpeechServiceConnection_RecoMode */ SpeechServiceConnection_RecoMode = 14, /** * The spoken language to be recognized (in BCP-47 format). Under normal circumstances, you shouldn't have to use this property * directly. * Instead, use [[SpeechConfig.speechRecognitionLanguage]]. * @member PropertyId.SpeechServiceConnection_RecoLanguage */ SpeechServiceConnection_RecoLanguage = 15, /** * The session id. This id is a universally unique identifier (aka UUID) representing a specific binding of an audio input stream * and the underlying speech recognition instance to which it is bound. Under normal circumstances, you shouldn't have to use this * property directly. * Instead use [[SessionEventArgs.sessionId]]. * @member PropertyId.Speech_SessionId */ Speech_SessionId = 16, /** * The spoken language to be synthesized (e.g. en-US) * @member PropertyId.SpeechServiceConnection_SynthLanguage */ SpeechServiceConnection_SynthLanguage = 17, /** * The name of the TTS voice to be used for speech synthesis * @member PropertyId.SpeechServiceConnection_SynthVoice */ SpeechServiceConnection_SynthVoice = 18, /** * The string to specify TTS output audio format * @member PropertyId.SpeechServiceConnection_SynthOutputFormat */ SpeechServiceConnection_SynthOutputFormat = 19, /** * The list of comma separated languages used as possible source languages * Added in version 1.13.0 * @member PropertyId.SpeechServiceConnection_AutoDetectSourceLanguages */ SpeechServiceConnection_AutoDetectSourceLanguages = 20, /** * The requested Cognitive Services Speech Service response output format (simple or detailed). Under normal circumstances, you shouldn't have * to use this property directly. * Instead use [[SpeechConfig.outputFormat]]. * @member PropertyId.SpeechServiceResponse_RequestDetailedResultTrueFalse */ SpeechServiceResponse_RequestDetailedResultTrueFalse = 21, /** * The requested Cognitive Services Speech Service response output profanity level. Currently unused. * @member PropertyId.SpeechServiceResponse_RequestProfanityFilterTrueFalse */ SpeechServiceResponse_RequestProfanityFilterTrueFalse = 22, /** * The Cognitive Services Speech Service response output (in JSON format). This property is available on recognition result objects only. * @member PropertyId.SpeechServiceResponse_JsonResult */ SpeechServiceResponse_JsonResult = 23, /** * The Cognitive Services Speech Service error details (in JSON format). Under normal circumstances, you shouldn't have to * use this property directly. Instead use [[CancellationDetails.errorDetails]]. * @member PropertyId.SpeechServiceResponse_JsonErrorDetails */ SpeechServiceResponse_JsonErrorDetails = 24, /** * The cancellation reason. Currently unused. * @member PropertyId.CancellationDetails_Reason */ CancellationDetails_Reason = 25, /** * The cancellation text. Currently unused. * @member PropertyId.CancellationDetails_ReasonText */ CancellationDetails_ReasonText = 26, /** * The Cancellation detailed text. Currently unused. * @member PropertyId.CancellationDetails_ReasonDetailedText */ CancellationDetails_ReasonDetailedText = 27, /** * The Language Understanding Service response output (in JSON format). Available via [[IntentRecognitionResult]] * @member PropertyId.LanguageUnderstandingServiceResponse_JsonResult */ LanguageUnderstandingServiceResponse_JsonResult = 28, /** * The URL string built from speech configuration. * This property is intended to be read-only. The SDK is using it internally. * NOTE: Added in version 1.7.0. */ SpeechServiceConnection_Url = 29, /** * The initial silence timeout value (in milliseconds) used by the service. * Added in version 1.7.0 */ SpeechServiceConnection_InitialSilenceTimeoutMs = 30, /** * The end silence timeout value (in milliseconds) used by the service. * Added in version 1.7.0 */ SpeechServiceConnection_EndSilenceTimeoutMs = 31, /** * A duration of detected silence, measured in milliseconds, after which speech-to-text will determine a spoken * phrase has ended and generate a final Recognized result. Configuring this timeout may be helpful in situations * where spoken input is significantly faster or slower than usual and default segmentation behavior consistently * yields results that are too long or too short. Segmentation timeout values that are inappropriately high or low * can negatively affect speech-to-text accuracy; this property should be carefully configured and the resulting * behavior should be thoroughly validated as intended. * * For more information about timeout configuration that includes discussion of default behaviors, please visit * https://aka.ms/csspeech/timeouts. * * Added in version 1.21.0. */ Speech_SegmentationSilenceTimeoutMs = 32, /** * SegmentationMaximumTimeMs represents the maximum length of a spoken phrase when using the Time segmentation strategy. * As the length of a spoken phrase approaches this value, the @member Speech_SegmentationSilenceTimeoutMs will be reduced until either * the phrase silence timeout is reached or the phrase reaches the maximum length. * * Added in version 1.42.0. */ Speech_SegmentationMaximumTimeMs = 33, /** * SegmentationStrategy defines the strategy used to determine when a spoken phrase has ended and a final Recognized result should be generated. * Allowed values are "Default", "Time", and "Semantic". * * Valid values: * - "Default": Uses the default strategy and settings as determined by the Speech Service. Suitable for most situations. * - "Time": Uses a time-based strategy where the amount of silence between speech determines when to generate a final result. * - "Semantic": Uses an AI model to determine the end of a spoken phrase based on the phrase's content. * * Additional Notes: * - When using the Time strategy, @member Speech_SegmentationSilenceTimeoutMs can be adjusted to modify the required silence duration for ending a phrase, * and @member Speech_SegmentationMaximumTimeMs can be adjusted to set the maximum length of a spoken phrase. * - The Semantic strategy does not have any adjustable properties. * * Added in version 1.42.0. */ Speech_SegmentationStrategy = 34, /** * A boolean value specifying whether audio logging is enabled in the service or not. * Audio and content logs are stored either in Microsoft-owned storage, or in your own storage account linked * to your Cognitive Services subscription (Bring Your Own Storage (BYOS) enabled Speech resource). * The logs will be removed after 30 days. * Added in version 1.7.0 */ SpeechServiceConnection_EnableAudioLogging = 35, /** * The speech service connection language identifier mode. * Can be "AtStart" (the default), or "Continuous". See Language * Identification document https://aka.ms/speech/lid?pivots=programming-language-javascript * for more details. * Added in 1.25.0 **/ SpeechServiceConnection_LanguageIdMode = 36, /** * A string value representing the desired endpoint version to target for Speech Recognition. * Added in version 1.21.0 */ SpeechServiceConnection_RecognitionEndpointVersion = 37, /** /** * A string value the current speaker recognition scenario/mode (TextIndependentIdentification, etc.). * Added in version 1.23.0 */ SpeechServiceConnection_SpeakerIdMode = 38, /** * The requested Cognitive Services Speech Service response output profanity setting. * Allowed values are "masked", "removed", and "raw". * Added in version 1.7.0. */ SpeechServiceResponse_ProfanityOption = 39, /** * A string value specifying which post processing option should be used by service. * Allowed values are "TrueText". * Added in version 1.7.0 */ SpeechServiceResponse_PostProcessingOption = 40, /** * A boolean value specifying whether to include word-level timestamps in the response result. * Added in version 1.7.0 */ SpeechServiceResponse_RequestWordLevelTimestamps = 41, /** * The number of times a word has to be in partial results to be returned. * Added in version 1.7.0 */ SpeechServiceResponse_StablePartialResultThreshold = 42, /** * A string value specifying the output format option in the response result. Internal use only. * Added in version 1.7.0. */ SpeechServiceResponse_OutputFormatOption = 43, /** * A boolean value to request for stabilizing translation partial results by omitting words in the end. * Added in version 1.7.0. */ SpeechServiceResponse_TranslationRequestStablePartialResult = 44, /** * A boolean value specifying whether to request WordBoundary events. * @member PropertyId.SpeechServiceResponse_RequestWordBoundary * Added in version 1.21.0. */ SpeechServiceResponse_RequestWordBoundary = 45, /** * A boolean value specifying whether to request punctuation boundary in WordBoundary Events. Default is true. * @member PropertyId.SpeechServiceResponse_RequestPunctuationBoundary * Added in version 1.21.0. */ SpeechServiceResponse_RequestPunctuationBoundary = 46, /** * A boolean value specifying whether to request sentence boundary in WordBoundary Events. Default is false. * @member PropertyId.SpeechServiceResponse_RequestSentenceBoundary * Added in version 1.21.0. */ SpeechServiceResponse_RequestSentenceBoundary = 47, /** * Determines if intermediate results contain speaker identification. * Allowed values are "true" or "false". If set to "true", the intermediate results will contain speaker identification. * The default value if unset or set to an invalid value is "false". * This is currently only supported for scenarios using the ConversationTranscriber". * @member PropertyId.SpeechServiceResponse_DiarizeIntermediateResults * Adding in version 1.41. */ SpeechServiceResponse_DiarizeIntermediateResults = 48, /** * Identifier used to connect to the backend service. * @member PropertyId.Conversation_ApplicationId */ Conversation_ApplicationId = 49, /** * Type of dialog backend to connect to. * @member PropertyId.Conversation_DialogType */ Conversation_DialogType = 50, /** * Silence timeout for listening * @member PropertyId.Conversation_Initial_Silence_Timeout */ Conversation_Initial_Silence_Timeout = 51, /** * From Id to add to speech recognition activities. * @member PropertyId.Conversation_From_Id */ Conversation_From_Id = 52, /** * ConversationId for the session. * @member PropertyId.Conversation_Conversation_Id */ Conversation_Conversation_Id = 53, /** * Comma separated list of custom voice deployment ids. * @member PropertyId.Conversation_Custom_Voice_Deployment_Ids */ Conversation_Custom_Voice_Deployment_Ids = 54, /** * Speech activity template, stamp properties from the template on the activity generated by the service for speech. * @member PropertyId.Conversation_Speech_Activity_Template * Added in version 1.10.0. */ Conversation_Speech_Activity_Template = 55, /** * Enables or disables the receipt of turn status messages as obtained on the turnStatusReceived event. * @member PropertyId.Conversation_Request_Bot_Status_Messages * Added in version 1.15.0. */ Conversation_Request_Bot_Status_Messages = 56, /** * Specifies the connection ID to be provided in the Agent configuration message, e.g. a Direct Line token for * channel authentication. * Added in version 1.15.1. */ Conversation_Agent_Connection_Id = 57, /** * The Cognitive Services Speech Service host (url). Under normal circumstances, you shouldn't have to use this property directly. * Instead, use [[SpeechConfig.fromHost]]. */ SpeechServiceConnection_Host = 58, /** * Set the host for service calls to the Conversation Translator REST management and websocket calls. */ ConversationTranslator_Host = 59, /** * Optionally set the the host's display name. * Used when joining a conversation. */ ConversationTranslator_Name = 60, /** * Optionally set a value for the X-CorrelationId request header. * Used for troubleshooting errors in the server logs. It should be a valid guid. */ ConversationTranslator_CorrelationId = 61, /** * Set the conversation token to be sent to the speech service. This enables the * service to service call from the speech service to the Conversation Translator service for relaying * recognitions. For internal use. */ ConversationTranslator_Token = 62, /** * The reference text of the audio for pronunciation evaluation. * For this and the following pronunciation assessment parameters, see * https://docs.microsoft.com/azure/cognitive-services/speech-service/rest-speech-to-text#pronunciation-assessment-parameters for details. * Under normal circumstances, you shouldn't have to use this property directly. * Added in version 1.15.0 */ PronunciationAssessment_ReferenceText = 63, /** * The point system for pronunciation score calibration (FivePoint or HundredMark). * Under normal circumstances, you shouldn't have to use this property directly. * Added in version 1.15.0 */ PronunciationAssessment_GradingSystem = 64, /** * The pronunciation evaluation granularity (Phoneme, Word, or FullText). * Under normal circumstances, you shouldn't have to use this property directly. * Added in version 1.15.0 */ PronunciationAssessment_Granularity = 65, /** * Defines if enable miscue calculation. * With this enabled, the pronounced words will be compared to the reference text, * and will be marked with omission/insertion based on the comparison. The default setting is False. * Under normal circumstances, you shouldn't have to use this property directly. * Added in version 1.15.0 */ PronunciationAssessment_EnableMiscue = 66, /** * The json string of pronunciation assessment parameters * Under normal circumstances, you shouldn't have to use this property directly. * Added in version 1.15.0 */ PronunciationAssessment_Json = 67, /** * Pronunciation assessment parameters. * This property is intended to be read-only. The SDK is using it internally. * Added in version 1.15.0 */ PronunciationAssessment_Params = 68, /** * Version of Speaker Recognition API to use. * Added in version 1.18.0 */ SpeakerRecognition_Api_Version = 69, /** * Specifies whether to allow load of data URL for web worker * Allowed values are "off" and "on". Default is "on". * Added in version 1.32.0 */ WebWorkerLoadType = 70, /** * Talking avatar service WebRTC session description protocol. * This property is intended to be read-only. The SDK is using it internally. * Added in version 1.33.0 */ TalkingAvatarService_WebRTC_SDP = 71 }