{"version":3,"sources":["src/sdk/PropertyId.ts"],"names":[],"mappings":"AAGA;;;GAGG;AACH,oBAAY,UAAU;IAElB;;;;;;OAMG;IACH,2BAA2B,IAAI;IAE/B;;;;;;OAMG;IACH,gCAAgC,IAAA;IAEhC;;;;;OAKG;IACH,8BAA8B,IAAA;IAE9B;;;;;;OAMG;IACH,gCAAgC,IAAA;IAEhC;;;OAGG;IACH,+BAA+B,IAAA;IAE/B;;;;;;OAMG;IACH,kCAAkC,IAAA;IAElC;;;;;;OAMG;IACH,8CAA8C,IAAA;IAE9C;;;;;;OAMG;IACH,wCAAwC,IAAA;IAExC;;;OAGG;IACH,2CAA2C,IAAA;IAE3C;;;;OAIG;IACH,oCAAoC,IAAA;IAEpC;;;;;OAKG;IACH,qCAAqC,KAAA;IAErC;;;;;OAKG;IACH,iCAAiC,KAAA;IAEjC;;;;;OAKG;IACH,qCAAqC,KAAA;IAErC;;;;;OAKG;IACH,qCAAqC,KAAA;IAErC;;;;OAIG;IACH,gCAAgC,KAAA;IAEhC;;;;;OAKG;IACH,oCAAoC,KAAA;IAEpC;;;;;;OAMG;IACH,gBAAgB,KAAA;IAEhB;;;;;OAKG;IACH,oDAAoD,KAAA;IAEpD;;;OAGG;IACH,qDAAqD,KAAA;IAErD;;;OAGG;IACH,gCAAgC,KAAA;IAEhC;;;;OAIG;IACH,sCAAsC,KAAA;IAEtC;;;OAGG;IACH,0BAA0B,KAAA;IAE1B;;;OAGG;IACH,8BAA8B,KAAA;IAE9B;;;OAGG;IACH,sCAAsC,KAAA;IAEtC;;;OAGG;IACH,+CAA+C,KAAA;IAEhD;;;;OAIG;IACF,2BAA2B,KAAA;IAE5B;;;OAGG;IACF,+CAA+C,KAAA;IAEhD;;;OAGG;IACF,2CAA2C,KAAA;IAE5C;;;OAGG;IACF,0CAA0C,KAAA;IAE3C;;;;OAIG;IACF,qCAAqC,KAAA;IAEtC;;;;OAIG;IACF,0CAA0C,KAAA;IAE3C;;;OAGG;IACF,gDAAgD,KAAA;IAEjD;;;OAGG;IACF,kDAAkD,KAAA;IAEnD;;;OAGG;IACF,wCAAwC,KAAA;IAEzC;;;OAGG;IACF,2DAA2D,KAAA;IAE3D;;;OAGG;IACH,0BAA0B,KAAA;IAE1B;;;OAGG;IACH,uBAAuB,KAAA;IAEvB;;;OAGG;IACH,oCAAoC,KAAA;IAEpC;;;OAGG;IACH,oBAAoB,KAAA;CACvB","file":"PropertyId.d.ts","sourcesContent":["// Copyright (c) Microsoft Corporation. All rights reserved.\n// Licensed under the MIT license.\n\n/**\n * Defines speech property ids.\n * @class PropertyId\n */\nexport enum PropertyId {\n\n    /**\n     * The Cognitive Services Speech Service subscription Key. If you are using an intent recognizer, you need to specify\n     * to specify the LUIS endpoint key for your particular LUIS app. Under normal circumstances, you shouldn't\n     * have to use this property directly.\n     * Instead, use [[SpeechConfig.fromSubscription]].\n     * @member PropertyId.SpeechServiceConnection_Key\n     */\n    SpeechServiceConnection_Key = 0,\n\n    /**\n     * The Cognitive Services Speech Service endpoint (url). Under normal circumstances, you shouldn't\n     * have to use this property directly.\n     * Instead, use [[SpeechConfig.fromEndpoint]].\n     * NOTE: This endpoint is not the same as the endpoint used to obtain an access token.\n     * @member PropertyId.SpeechServiceConnection_Endpoint\n     */\n    SpeechServiceConnection_Endpoint,\n\n    /**\n     * The Cognitive Services Speech Service region. Under normal circumstances, you shouldn't have to\n     * use this property directly.\n     * Instead, use [[SpeechConfig.fromSubscription]], [[SpeechConfig.fromEndpoint]], [[SpeechConfig.fromAuthorizationToken]].\n     * @member PropertyId.SpeechServiceConnection_Region\n     */\n    SpeechServiceConnection_Region,\n\n    /**\n     * The Cognitive Services Speech Service authorization token (aka access token). Under normal circumstances,\n     * you shouldn't have to use this property directly.\n     * Instead, use [[SpeechConfig.fromAuthorizationToken]],\n     * [[SpeechRecognizer.authorizationToken]], [[IntentRecognizer.authorizationToken]], [[TranslationRecognizer.authorizationToken]].\n     * @member PropertyId.SpeechServiceAuthorization_Token\n     */\n    SpeechServiceAuthorization_Token,\n\n    /**\n     * The Cognitive Services Speech Service authorization type. Currently unused.\n     * @member PropertyId.SpeechServiceAuthorization_Type\n     */\n    SpeechServiceAuthorization_Type,\n\n    /**\n     * The Cognitive Services Speech Service endpoint id. Under normal circumstances, you shouldn't\n     * have to use this property directly.\n     * Instead, use [[SpeechConfig.endpointId]].\n     * NOTE: The endpoint id is available in the Speech Portal, listed under Endpoint Details.\n     * @member PropertyId.SpeechServiceConnection_EndpointId\n     */\n    SpeechServiceConnection_EndpointId,\n\n    /**\n     * The list of comma separated languages (BCP-47 format) used as target translation languages. Under normal circumstances,\n     * you shouldn't have to use this property directly.\n     * Instead use [[SpeechTranslationConfig.addTargetLanguage]],\n     * [[SpeechTranslationConfig.targetLanguages]], [[TranslationRecognizer.targetLanguages]].\n     * @member PropertyId.SpeechServiceConnection_TranslationToLanguages\n     */\n    SpeechServiceConnection_TranslationToLanguages,\n\n    /**\n     * The name of the Cognitive Service Text to Speech Service Voice. Under normal circumstances, you shouldn't have to use this\n     * property directly.\n     * Instead, use [[SpeechTranslationConfig.voiceName]].\n     * NOTE: Valid voice names can be found <a href=\"https://aka.ms/csspeech/voicenames\">here</a>.\n     * @member PropertyId.SpeechServiceConnection_TranslationVoice\n     */\n    SpeechServiceConnection_TranslationVoice,\n\n    /**\n     * Translation features.\n     * @member PropertyId.SpeechServiceConnection_TranslationFeatures\n     */\n    SpeechServiceConnection_TranslationFeatures,\n\n    /**\n     * The Language Understanding Service Region. Under normal circumstances, you shouldn't have to use this property directly.\n     * Instead, use [[LanguageUnderstandingModel]].\n     * @member PropertyId.SpeechServiceConnection_IntentRegion\n     */\n    SpeechServiceConnection_IntentRegion,\n\n    /**\n     * The host name of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments.\n     * You shouldn't have to use this property directly.\n     * Instead use <see cref=\"SpeechConfig.SetProxy(string,int,string,string)\"/>.\n     * Added in version 1.4.0.\n     */\n    SpeechServiceConnection_ProxyHostName,\n\n    /**\n     * The port of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments.\n     * You shouldn't have to use this property directly.\n     * Instead use <see cref=\"SpeechConfig.SetProxy(string,int,string,string)\"/>.\n     * Added in version 1.4.0.\n     */\n    SpeechServiceConnection_ProxyPort,\n\n    /**\n     * The user name of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments.\n     * You shouldn't have to use this property directly.\n     * Instead use <see cref=\"SpeechConfig.SetProxy(string,int,string,string)\"/>.\n     * Added in version 1.4.0.\n     */\n    SpeechServiceConnection_ProxyUserName,\n\n    /**\n     * The password of the proxy server used to connect to the Cognitive Services Speech Service. Only relevant in Node.js environments.\n     * You shouldn't have to use this property directly.\n     * Instead use <see cref=\"SpeechConfig.SetProxy(string,int,string,string)\"/>.\n     * Added in version 1.4.0.\n     */\n    SpeechServiceConnection_ProxyPassword,\n\n    /**\n     * The Cognitive Services Speech Service recognition Mode. Can be \"INTERACTIVE\", \"CONVERSATION\", \"DICTATION\".\n     * This property is intended to be read-only. The SDK is using it internally.\n     * @member PropertyId.SpeechServiceConnection_RecoMode\n     */\n    SpeechServiceConnection_RecoMode,\n\n    /**\n     * The spoken language to be recognized (in BCP-47 format). Under normal circumstances, you shouldn't have to use this property\n     * directly.\n     * Instead, use [[SpeechConfig.speechRecognitionLanguage]].\n     * @member PropertyId.SpeechServiceConnection_RecoLanguage\n     */\n    SpeechServiceConnection_RecoLanguage,\n\n    /**\n     * The session id. This id is a universally unique identifier (aka UUID) representing a specific binding of an audio input stream\n     * and the underlying speech recognition instance to which it is bound. Under normal circumstances, you shouldn't have to use this\n     * property directly.\n     * Instead use [[SessionEventArgs.sessionId]].\n     * @member PropertyId.Speech_SessionId\n     */\n    Speech_SessionId,\n\n    /**\n     * The requested Cognitive Services Speech Service response output format (simple or detailed). Under normal circumstances, you shouldn't have\n     * to use this property directly.\n     * Instead use [[SpeechConfig.outputFormat]].\n     * @member PropertyId.SpeechServiceResponse_RequestDetailedResultTrueFalse\n     */\n    SpeechServiceResponse_RequestDetailedResultTrueFalse,\n\n    /**\n     * The requested Cognitive Services Speech Service response output profanity level. Currently unused.\n     * @member PropertyId.SpeechServiceResponse_RequestProfanityFilterTrueFalse\n     */\n    SpeechServiceResponse_RequestProfanityFilterTrueFalse,\n\n    /**\n     * The Cognitive Services Speech Service response output (in JSON format). This property is available on recognition result objects only.\n     * @member PropertyId.SpeechServiceResponse_JsonResult\n     */\n    SpeechServiceResponse_JsonResult,\n\n    /**\n     * The Cognitive Services Speech Service error details (in JSON format). Under normal circumstances, you shouldn't have to\n     * use this property directly. Instead use [[CancellationDetails.errorDetails]].\n     * @member PropertyId.SpeechServiceResponse_JsonErrorDetails\n     */\n    SpeechServiceResponse_JsonErrorDetails,\n\n    /**\n     * The cancellation reason. Currently unused.\n     * @member PropertyId.CancellationDetails_Reason\n     */\n    CancellationDetails_Reason,\n\n    /**\n     * The cancellation text. Currently unused.\n     * @member PropertyId.CancellationDetails_ReasonText\n     */\n    CancellationDetails_ReasonText,\n\n    /**\n     * The Cancellation detailed text. Currently unused.\n     * @member PropertyId.CancellationDetails_ReasonDetailedText\n     */\n    CancellationDetails_ReasonDetailedText,\n\n    /**\n     * The Language Understanding Service response output (in JSON format). Available via [[IntentRecognitionResult]]\n     * @member PropertyId.LanguageUnderstandingServiceResponse_JsonResult\n     */\n    LanguageUnderstandingServiceResponse_JsonResult,\n\n   /**\n    * The URL string built from speech configuration.\n    * This property is intended to be read-only. The SDK is using it internally.\n    * NOTE: Added in version 1.7.0.\n    */\n    SpeechServiceConnection_Url,\n\n   /**\n    * The initial silence timeout value (in milliseconds) used by the service.\n    * Added in version 1.7.0\n    */\n    SpeechServiceConnection_InitialSilenceTimeoutMs,\n\n   /**\n    * The end silence timeout value (in milliseconds) used by the service.\n    * Added in version 1.7.0\n    */\n    SpeechServiceConnection_EndSilenceTimeoutMs,\n\n   /**\n    * A boolean value specifying whether audio logging is enabled in the service or not.\n    * Added in version 1.7.0\n    */\n    SpeechServiceConnection_EnableAudioLogging,\n\n   /**\n    * The requested Cognitive Services Speech Service response output profanity setting.\n    * Allowed values are \"masked\", \"removed\", and \"raw\".\n    * Added in version 1.7.0.\n    */\n    SpeechServiceResponse_ProfanityOption,\n\n   /**\n    * A string value specifying which post processing option should be used by service.\n    * Allowed values are \"TrueText\".\n    * Added in version 1.7.0\n    */\n    SpeechServiceResponse_PostProcessingOption,\n\n   /**\n    *  A boolean value specifying whether to include word-level timestamps in the response result.\n    * Added in version 1.7.0\n    */\n    SpeechServiceResponse_RequestWordLevelTimestamps,\n\n   /**\n    * The number of times a word has to be in partial results to be returned.\n    * Added in version 1.7.0\n    */\n    SpeechServiceResponse_StablePartialResultThreshold,\n\n   /**\n    * A string value specifying the output format option in the response result. Internal use only.\n    * Added in version 1.7.0.\n    */\n    SpeechServiceResponse_OutputFormatOption,\n\n   /**\n    * A boolean value to request for stabilizing translation partial results by omitting words in the end.\n    * Added in version 1.7.0.\n    */\n    SpeechServiceResponse_TranslationRequestStablePartialResult,\n\n    /**\n     * Identifier used to connect to the backend service.\n     * @member PropertyId.Conversation_ApplicationId\n     */\n    Conversation_ApplicationId,\n\n    /**\n     * Type of dialog backend to connect to.\n     * @member PropertyId.Conversation_DialogType\n     */\n    Conversation_DialogType,\n\n    /**\n     * Silence timeout for listening\n     * @member PropertyId.Conversation_Initial_Silence_Timeout\n     */\n    Conversation_Initial_Silence_Timeout,\n\n    /**\n     * From Id to add to speech recognition activities.\n     * @member PropertyId.Conversation_From_Id\n     */\n    Conversation_From_Id\n}\n"]}