universal-s3
Version:
Universal S3 SDK for JavaScript, available for Node.js backends
728 lines • 279 kB
TypeScript
import {Request} from '../lib/request';
import {Response} from '../lib/response';
import {AWSError} from '../lib/error';
import {Service} from '../lib/service';
import {ServiceConfigurationOptions} from '../lib/service';
import {ConfigBase as Config} from '../lib/config';
interface Blob {}
declare class MediaConvert extends Service {
/**
* Constructs a service object. This object has one method for each API operation.
*/
constructor(options?: MediaConvert.Types.ClientConfiguration)
config: Config & MediaConvert.Types.ClientConfiguration;
/**
* Associates an AWS Certificate Manager (ACM) Amazon Resource Name (ARN) with AWS Elemental MediaConvert.
*/
associateCertificate(params: MediaConvert.Types.AssociateCertificateRequest, callback?: (err: AWSError, data: MediaConvert.Types.AssociateCertificateResponse) => void): Request<MediaConvert.Types.AssociateCertificateResponse, AWSError>;
/**
* Associates an AWS Certificate Manager (ACM) Amazon Resource Name (ARN) with AWS Elemental MediaConvert.
*/
associateCertificate(callback?: (err: AWSError, data: MediaConvert.Types.AssociateCertificateResponse) => void): Request<MediaConvert.Types.AssociateCertificateResponse, AWSError>;
/**
* Permanently cancel a job. Once you have canceled a job, you can't start it again.
*/
cancelJob(params: MediaConvert.Types.CancelJobRequest, callback?: (err: AWSError, data: MediaConvert.Types.CancelJobResponse) => void): Request<MediaConvert.Types.CancelJobResponse, AWSError>;
/**
* Permanently cancel a job. Once you have canceled a job, you can't start it again.
*/
cancelJob(callback?: (err: AWSError, data: MediaConvert.Types.CancelJobResponse) => void): Request<MediaConvert.Types.CancelJobResponse, AWSError>;
/**
* Create a new transcoding job. For information about jobs and job settings, see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createJob(params: MediaConvert.Types.CreateJobRequest, callback?: (err: AWSError, data: MediaConvert.Types.CreateJobResponse) => void): Request<MediaConvert.Types.CreateJobResponse, AWSError>;
/**
* Create a new transcoding job. For information about jobs and job settings, see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createJob(callback?: (err: AWSError, data: MediaConvert.Types.CreateJobResponse) => void): Request<MediaConvert.Types.CreateJobResponse, AWSError>;
/**
* Create a new job template. For information about job templates see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createJobTemplate(params: MediaConvert.Types.CreateJobTemplateRequest, callback?: (err: AWSError, data: MediaConvert.Types.CreateJobTemplateResponse) => void): Request<MediaConvert.Types.CreateJobTemplateResponse, AWSError>;
/**
* Create a new job template. For information about job templates see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createJobTemplate(callback?: (err: AWSError, data: MediaConvert.Types.CreateJobTemplateResponse) => void): Request<MediaConvert.Types.CreateJobTemplateResponse, AWSError>;
/**
* Create a new preset. For information about job templates see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createPreset(params: MediaConvert.Types.CreatePresetRequest, callback?: (err: AWSError, data: MediaConvert.Types.CreatePresetResponse) => void): Request<MediaConvert.Types.CreatePresetResponse, AWSError>;
/**
* Create a new preset. For information about job templates see the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
*/
createPreset(callback?: (err: AWSError, data: MediaConvert.Types.CreatePresetResponse) => void): Request<MediaConvert.Types.CreatePresetResponse, AWSError>;
/**
* Create a new transcoding queue. For information about queues, see Working With Queues in the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-queues.html
*/
createQueue(params: MediaConvert.Types.CreateQueueRequest, callback?: (err: AWSError, data: MediaConvert.Types.CreateQueueResponse) => void): Request<MediaConvert.Types.CreateQueueResponse, AWSError>;
/**
* Create a new transcoding queue. For information about queues, see Working With Queues in the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-queues.html
*/
createQueue(callback?: (err: AWSError, data: MediaConvert.Types.CreateQueueResponse) => void): Request<MediaConvert.Types.CreateQueueResponse, AWSError>;
/**
* Permanently delete a job template you have created.
*/
deleteJobTemplate(params: MediaConvert.Types.DeleteJobTemplateRequest, callback?: (err: AWSError, data: MediaConvert.Types.DeleteJobTemplateResponse) => void): Request<MediaConvert.Types.DeleteJobTemplateResponse, AWSError>;
/**
* Permanently delete a job template you have created.
*/
deleteJobTemplate(callback?: (err: AWSError, data: MediaConvert.Types.DeleteJobTemplateResponse) => void): Request<MediaConvert.Types.DeleteJobTemplateResponse, AWSError>;
/**
* Permanently delete a preset you have created.
*/
deletePreset(params: MediaConvert.Types.DeletePresetRequest, callback?: (err: AWSError, data: MediaConvert.Types.DeletePresetResponse) => void): Request<MediaConvert.Types.DeletePresetResponse, AWSError>;
/**
* Permanently delete a preset you have created.
*/
deletePreset(callback?: (err: AWSError, data: MediaConvert.Types.DeletePresetResponse) => void): Request<MediaConvert.Types.DeletePresetResponse, AWSError>;
/**
* Permanently delete a queue you have created.
*/
deleteQueue(params: MediaConvert.Types.DeleteQueueRequest, callback?: (err: AWSError, data: MediaConvert.Types.DeleteQueueResponse) => void): Request<MediaConvert.Types.DeleteQueueResponse, AWSError>;
/**
* Permanently delete a queue you have created.
*/
deleteQueue(callback?: (err: AWSError, data: MediaConvert.Types.DeleteQueueResponse) => void): Request<MediaConvert.Types.DeleteQueueResponse, AWSError>;
/**
* Send an request with an empty body to the regional API endpoint to get your account API endpoint.
*/
describeEndpoints(params: MediaConvert.Types.DescribeEndpointsRequest, callback?: (err: AWSError, data: MediaConvert.Types.DescribeEndpointsResponse) => void): Request<MediaConvert.Types.DescribeEndpointsResponse, AWSError>;
/**
* Send an request with an empty body to the regional API endpoint to get your account API endpoint.
*/
describeEndpoints(callback?: (err: AWSError, data: MediaConvert.Types.DescribeEndpointsResponse) => void): Request<MediaConvert.Types.DescribeEndpointsResponse, AWSError>;
/**
* Removes an association between the Amazon Resource Name (ARN) of an AWS Certificate Manager (ACM) certificate and an AWS Elemental MediaConvert resource.
*/
disassociateCertificate(params: MediaConvert.Types.DisassociateCertificateRequest, callback?: (err: AWSError, data: MediaConvert.Types.DisassociateCertificateResponse) => void): Request<MediaConvert.Types.DisassociateCertificateResponse, AWSError>;
/**
* Removes an association between the Amazon Resource Name (ARN) of an AWS Certificate Manager (ACM) certificate and an AWS Elemental MediaConvert resource.
*/
disassociateCertificate(callback?: (err: AWSError, data: MediaConvert.Types.DisassociateCertificateResponse) => void): Request<MediaConvert.Types.DisassociateCertificateResponse, AWSError>;
/**
* Retrieve the JSON for a specific completed transcoding job.
*/
getJob(params: MediaConvert.Types.GetJobRequest, callback?: (err: AWSError, data: MediaConvert.Types.GetJobResponse) => void): Request<MediaConvert.Types.GetJobResponse, AWSError>;
/**
* Retrieve the JSON for a specific completed transcoding job.
*/
getJob(callback?: (err: AWSError, data: MediaConvert.Types.GetJobResponse) => void): Request<MediaConvert.Types.GetJobResponse, AWSError>;
/**
* Retrieve the JSON for a specific job template.
*/
getJobTemplate(params: MediaConvert.Types.GetJobTemplateRequest, callback?: (err: AWSError, data: MediaConvert.Types.GetJobTemplateResponse) => void): Request<MediaConvert.Types.GetJobTemplateResponse, AWSError>;
/**
* Retrieve the JSON for a specific job template.
*/
getJobTemplate(callback?: (err: AWSError, data: MediaConvert.Types.GetJobTemplateResponse) => void): Request<MediaConvert.Types.GetJobTemplateResponse, AWSError>;
/**
* Retrieve the JSON for a specific preset.
*/
getPreset(params: MediaConvert.Types.GetPresetRequest, callback?: (err: AWSError, data: MediaConvert.Types.GetPresetResponse) => void): Request<MediaConvert.Types.GetPresetResponse, AWSError>;
/**
* Retrieve the JSON for a specific preset.
*/
getPreset(callback?: (err: AWSError, data: MediaConvert.Types.GetPresetResponse) => void): Request<MediaConvert.Types.GetPresetResponse, AWSError>;
/**
* Retrieve the JSON for a specific queue.
*/
getQueue(params: MediaConvert.Types.GetQueueRequest, callback?: (err: AWSError, data: MediaConvert.Types.GetQueueResponse) => void): Request<MediaConvert.Types.GetQueueResponse, AWSError>;
/**
* Retrieve the JSON for a specific queue.
*/
getQueue(callback?: (err: AWSError, data: MediaConvert.Types.GetQueueResponse) => void): Request<MediaConvert.Types.GetQueueResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your job templates. This will return the templates themselves, not just a list of them. To retrieve the next twenty templates, use the nextToken string returned with the array
*/
listJobTemplates(params: MediaConvert.Types.ListJobTemplatesRequest, callback?: (err: AWSError, data: MediaConvert.Types.ListJobTemplatesResponse) => void): Request<MediaConvert.Types.ListJobTemplatesResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your job templates. This will return the templates themselves, not just a list of them. To retrieve the next twenty templates, use the nextToken string returned with the array
*/
listJobTemplates(callback?: (err: AWSError, data: MediaConvert.Types.ListJobTemplatesResponse) => void): Request<MediaConvert.Types.ListJobTemplatesResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your most recently created jobs. This array includes in-process, completed, and errored jobs. This will return the jobs themselves, not just a list of the jobs. To retrieve the twenty next most recent jobs, use the nextToken string returned with the array.
*/
listJobs(params: MediaConvert.Types.ListJobsRequest, callback?: (err: AWSError, data: MediaConvert.Types.ListJobsResponse) => void): Request<MediaConvert.Types.ListJobsResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your most recently created jobs. This array includes in-process, completed, and errored jobs. This will return the jobs themselves, not just a list of the jobs. To retrieve the twenty next most recent jobs, use the nextToken string returned with the array.
*/
listJobs(callback?: (err: AWSError, data: MediaConvert.Types.ListJobsResponse) => void): Request<MediaConvert.Types.ListJobsResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your presets. This will return the presets themselves, not just a list of them. To retrieve the next twenty presets, use the nextToken string returned with the array.
*/
listPresets(params: MediaConvert.Types.ListPresetsRequest, callback?: (err: AWSError, data: MediaConvert.Types.ListPresetsResponse) => void): Request<MediaConvert.Types.ListPresetsResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your presets. This will return the presets themselves, not just a list of them. To retrieve the next twenty presets, use the nextToken string returned with the array.
*/
listPresets(callback?: (err: AWSError, data: MediaConvert.Types.ListPresetsResponse) => void): Request<MediaConvert.Types.ListPresetsResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your queues. This will return the queues themselves, not just a list of them. To retrieve the next twenty queues, use the nextToken string returned with the array.
*/
listQueues(params: MediaConvert.Types.ListQueuesRequest, callback?: (err: AWSError, data: MediaConvert.Types.ListQueuesResponse) => void): Request<MediaConvert.Types.ListQueuesResponse, AWSError>;
/**
* Retrieve a JSON array of up to twenty of your queues. This will return the queues themselves, not just a list of them. To retrieve the next twenty queues, use the nextToken string returned with the array.
*/
listQueues(callback?: (err: AWSError, data: MediaConvert.Types.ListQueuesResponse) => void): Request<MediaConvert.Types.ListQueuesResponse, AWSError>;
/**
* Retrieve the tags for a MediaConvert resource.
*/
listTagsForResource(params: MediaConvert.Types.ListTagsForResourceRequest, callback?: (err: AWSError, data: MediaConvert.Types.ListTagsForResourceResponse) => void): Request<MediaConvert.Types.ListTagsForResourceResponse, AWSError>;
/**
* Retrieve the tags for a MediaConvert resource.
*/
listTagsForResource(callback?: (err: AWSError, data: MediaConvert.Types.ListTagsForResourceResponse) => void): Request<MediaConvert.Types.ListTagsForResourceResponse, AWSError>;
/**
* Add tags to a MediaConvert queue, preset, or job template. For information about tagging, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/tagging-resources.html
*/
tagResource(params: MediaConvert.Types.TagResourceRequest, callback?: (err: AWSError, data: MediaConvert.Types.TagResourceResponse) => void): Request<MediaConvert.Types.TagResourceResponse, AWSError>;
/**
* Add tags to a MediaConvert queue, preset, or job template. For information about tagging, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/tagging-resources.html
*/
tagResource(callback?: (err: AWSError, data: MediaConvert.Types.TagResourceResponse) => void): Request<MediaConvert.Types.TagResourceResponse, AWSError>;
/**
* Remove tags from a MediaConvert queue, preset, or job template. For information about tagging, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/tagging-resources.html
*/
untagResource(params: MediaConvert.Types.UntagResourceRequest, callback?: (err: AWSError, data: MediaConvert.Types.UntagResourceResponse) => void): Request<MediaConvert.Types.UntagResourceResponse, AWSError>;
/**
* Remove tags from a MediaConvert queue, preset, or job template. For information about tagging, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/tagging-resources.html
*/
untagResource(callback?: (err: AWSError, data: MediaConvert.Types.UntagResourceResponse) => void): Request<MediaConvert.Types.UntagResourceResponse, AWSError>;
/**
* Modify one of your existing job templates.
*/
updateJobTemplate(params: MediaConvert.Types.UpdateJobTemplateRequest, callback?: (err: AWSError, data: MediaConvert.Types.UpdateJobTemplateResponse) => void): Request<MediaConvert.Types.UpdateJobTemplateResponse, AWSError>;
/**
* Modify one of your existing job templates.
*/
updateJobTemplate(callback?: (err: AWSError, data: MediaConvert.Types.UpdateJobTemplateResponse) => void): Request<MediaConvert.Types.UpdateJobTemplateResponse, AWSError>;
/**
* Modify one of your existing presets.
*/
updatePreset(params: MediaConvert.Types.UpdatePresetRequest, callback?: (err: AWSError, data: MediaConvert.Types.UpdatePresetResponse) => void): Request<MediaConvert.Types.UpdatePresetResponse, AWSError>;
/**
* Modify one of your existing presets.
*/
updatePreset(callback?: (err: AWSError, data: MediaConvert.Types.UpdatePresetResponse) => void): Request<MediaConvert.Types.UpdatePresetResponse, AWSError>;
/**
* Modify one of your existing queues.
*/
updateQueue(params: MediaConvert.Types.UpdateQueueRequest, callback?: (err: AWSError, data: MediaConvert.Types.UpdateQueueResponse) => void): Request<MediaConvert.Types.UpdateQueueResponse, AWSError>;
/**
* Modify one of your existing queues.
*/
updateQueue(callback?: (err: AWSError, data: MediaConvert.Types.UpdateQueueResponse) => void): Request<MediaConvert.Types.UpdateQueueResponse, AWSError>;
}
declare namespace MediaConvert {
export type AacAudioDescriptionBroadcasterMix = "BROADCASTER_MIXED_AD"|"NORMAL"|string;
export type AacCodecProfile = "LC"|"HEV1"|"HEV2"|string;
export type AacCodingMode = "AD_RECEIVER_MIX"|"CODING_MODE_1_0"|"CODING_MODE_1_1"|"CODING_MODE_2_0"|"CODING_MODE_5_1"|string;
export type AacRateControlMode = "CBR"|"VBR"|string;
export type AacRawFormat = "LATM_LOAS"|"NONE"|string;
export interface AacSettings {
/**
* Choose BROADCASTER_MIXED_AD when the input contains pre-mixed main audio + audio description (AD) as a stereo pair. The value for AudioType will be set to 3, which signals to downstream systems that this stream contains "broadcaster mixed AD". Note that the input received by the encoder must contain pre-mixed audio; the encoder does not perform the mixing. When you choose BROADCASTER_MIXED_AD, the encoder ignores any values you provide in AudioType and FollowInputAudioType. Choose NORMAL when the input does not contain pre-mixed audio + audio description (AD). In this case, the encoder will use any values you provide for AudioType and FollowInputAudioType.
*/
AudioDescriptionBroadcasterMix?: AacAudioDescriptionBroadcasterMix;
/**
* Specify the average bitrate in bits per second. The set of valid values for this setting is: 6000, 8000, 10000, 12000, 14000, 16000, 20000, 24000, 28000, 32000, 40000, 48000, 56000, 64000, 80000, 96000, 112000, 128000, 160000, 192000, 224000, 256000, 288000, 320000, 384000, 448000, 512000, 576000, 640000, 768000, 896000, 1024000. The value you set is also constrained by the values that you choose for Profile (codecProfile), Bitrate control mode (codingMode), and Sample rate (sampleRate). Default values depend on Bitrate control mode and Profile.
*/
Bitrate?: __integerMin6000Max1024000;
/**
* AAC Profile.
*/
CodecProfile?: AacCodecProfile;
/**
* Mono (Audio Description), Mono, Stereo, or 5.1 channel layout. Valid values depend on rate control mode and profile. "1.0 - Audio Description (Receiver Mix)" setting receives a stereo description plus control track and emits a mono AAC encode of the description track, with control data emitted in the PES header as per ETSI TS 101 154 Annex E.
*/
CodingMode?: AacCodingMode;
/**
* Rate Control Mode.
*/
RateControlMode?: AacRateControlMode;
/**
* Enables LATM/LOAS AAC output. Note that if you use LATM/LOAS AAC in an output, you must choose "No container" for the output container.
*/
RawFormat?: AacRawFormat;
/**
* Sample rate in Hz. Valid values depend on rate control mode and profile.
*/
SampleRate?: __integerMin8000Max96000;
/**
* Use MPEG-2 AAC instead of MPEG-4 AAC audio for raw or MPEG-2 Transport Stream containers.
*/
Specification?: AacSpecification;
/**
* VBR Quality Level - Only used if rate_control_mode is VBR.
*/
VbrQuality?: AacVbrQuality;
}
export type AacSpecification = "MPEG2"|"MPEG4"|string;
export type AacVbrQuality = "LOW"|"MEDIUM_LOW"|"MEDIUM_HIGH"|"HIGH"|string;
export type Ac3BitstreamMode = "COMPLETE_MAIN"|"COMMENTARY"|"DIALOGUE"|"EMERGENCY"|"HEARING_IMPAIRED"|"MUSIC_AND_EFFECTS"|"VISUALLY_IMPAIRED"|"VOICE_OVER"|string;
export type Ac3CodingMode = "CODING_MODE_1_0"|"CODING_MODE_1_1"|"CODING_MODE_2_0"|"CODING_MODE_3_2_LFE"|string;
export type Ac3DynamicRangeCompressionProfile = "FILM_STANDARD"|"NONE"|string;
export type Ac3LfeFilter = "ENABLED"|"DISABLED"|string;
export type Ac3MetadataControl = "FOLLOW_INPUT"|"USE_CONFIGURED"|string;
export interface Ac3Settings {
/**
* Specify the average bitrate in bits per second. Valid bitrates depend on the coding mode.
*/
Bitrate?: __integerMin64000Max640000;
/**
* Specify the bitstream mode for the AC-3 stream that the encoder emits. For more information about the AC3 bitstream mode, see ATSC A/52-2012 (Annex E).
*/
BitstreamMode?: Ac3BitstreamMode;
/**
* Dolby Digital coding mode. Determines number of channels.
*/
CodingMode?: Ac3CodingMode;
/**
* Sets the dialnorm for the output. If blank and input audio is Dolby Digital, dialnorm will be passed through.
*/
Dialnorm?: __integerMin1Max31;
/**
* If set to FILM_STANDARD, adds dynamic range compression signaling to the output bitstream as defined in the Dolby Digital specification.
*/
DynamicRangeCompressionProfile?: Ac3DynamicRangeCompressionProfile;
/**
* Applies a 120Hz lowpass filter to the LFE channel prior to encoding. Only valid with 3_2_LFE coding mode.
*/
LfeFilter?: Ac3LfeFilter;
/**
* When set to FOLLOW_INPUT, encoder metadata will be sourced from the DD, DD+, or DolbyE decoder that supplied this audio data. If audio was not supplied from one of these streams, then the static metadata settings will be used.
*/
MetadataControl?: Ac3MetadataControl;
/**
* This value is always 48000. It represents the sample rate in Hz.
*/
SampleRate?: __integerMin48000Max48000;
}
export type AccelerationMode = "DISABLED"|"ENABLED"|"PREFERRED"|string;
export interface AccelerationSettings {
/**
* Specify the conditions when the service will run your job with accelerated transcoding.
*/
Mode: AccelerationMode;
}
export type AccelerationStatus = "NOT_APPLICABLE"|"IN_PROGRESS"|"ACCELERATED"|"NOT_ACCELERATED"|string;
export type AfdSignaling = "NONE"|"AUTO"|"FIXED"|string;
export interface AiffSettings {
/**
* Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track.
*/
BitDepth?: __integerMin16Max24;
/**
* Specify the number of channels in this output audio track. Valid values are 1 and even numbers up to 64. For example, 1, 2, 4, 6, and so on, up to 64.
*/
Channels?: __integerMin1Max64;
/**
* Sample rate in hz.
*/
SampleRate?: __integerMin8000Max192000;
}
export type AncillaryConvert608To708 = "UPCONVERT"|"DISABLED"|string;
export interface AncillarySourceSettings {
/**
* Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert (UPCONVERT), MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.
*/
Convert608To708?: AncillaryConvert608To708;
/**
* Specifies the 608 channel number in the ancillary data track from which to extract captions. Unused for passthrough.
*/
SourceAncillaryChannelNumber?: __integerMin1Max4;
/**
* By default, the service terminates any unterminated captions at the end of each input. If you want the caption to continue onto your next input, disable this setting.
*/
TerminateCaptions?: AncillaryTerminateCaptions;
}
export type AncillaryTerminateCaptions = "END_OF_INPUT"|"DISABLED"|string;
export type AntiAlias = "DISABLED"|"ENABLED"|string;
export interface AssociateCertificateRequest {
/**
* The ARN of the ACM certificate that you want to associate with your MediaConvert resource.
*/
Arn: __string;
}
export interface AssociateCertificateResponse {
}
export type AudioCodec = "AAC"|"MP2"|"WAV"|"AIFF"|"AC3"|"EAC3"|"EAC3_ATMOS"|"PASSTHROUGH"|string;
export interface AudioCodecSettings {
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode (rateControlMode) to "VBR" or "CBR". In VBR mode, you control the audio quality with the setting VBR quality (vbrQuality). In CBR mode, you use the setting Bitrate (bitrate). Defaults and valid values depend on the rate control mode.
*/
AacSettings?: AacSettings;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AC3.
*/
Ac3Settings?: Ac3Settings;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AIFF.
*/
AiffSettings?: AiffSettings;
/**
* Type of Audio codec.
*/
Codec?: AudioCodec;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3_ATMOS.
*/
Eac3AtmosSettings?: Eac3AtmosSettings;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value EAC3.
*/
Eac3Settings?: Eac3Settings;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value MP2.
*/
Mp2Settings?: Mp2Settings;
/**
* Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value WAV.
*/
WavSettings?: WavSettings;
}
export type AudioDefaultSelection = "DEFAULT"|"NOT_DEFAULT"|string;
export interface AudioDescription {
/**
* Advanced audio normalization settings. Ignore these settings unless you need to comply with a loudness standard.
*/
AudioNormalizationSettings?: AudioNormalizationSettings;
/**
* Specifies which audio data to use from each input. In the simplest case, specify an "Audio Selector":#inputs-audio_selector by name based on its order within each input. For example if you specify "Audio Selector 3", then the third audio selector will be used from each input. If an input does not have an "Audio Selector 3", then the audio selector marked as "default" in that input will be used. If there is no audio selector marked as "default", silence will be inserted for the duration of that input. Alternatively, an "Audio Selector Group":#inputs-audio_selector_group name may be specified, with similar default/silence behavior. If no audio_source_name is specified, then "Audio Selector 1" will be chosen automatically.
*/
AudioSourceName?: __string;
/**
* Applies only if Follow Input Audio Type is unchecked (false). A number between 0 and 255. The following are defined in ISO-IEC 13818-1: 0 = Undefined, 1 = Clean Effects, 2 = Hearing Impaired, 3 = Visually Impaired Commentary, 4-255 = Reserved.
*/
AudioType?: __integerMin0Max255;
/**
* When set to FOLLOW_INPUT, if the input contains an ISO 639 audio_type, then that value is passed through to the output. If the input contains no ISO 639 audio_type, the value in Audio Type is included in the output. Otherwise the value in Audio Type is included in the output. Note that this field and audioType are both ignored if audioDescriptionBroadcasterMix is set to BROADCASTER_MIXED_AD.
*/
AudioTypeControl?: AudioTypeControl;
/**
* Audio codec settings (CodecSettings) under (AudioDescriptions) contains the group of settings related to audio encoding. The settings in this group vary depending on the value that you choose for Audio codec (Codec). For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AAC, AacSettings * MP2, Mp2Settings * WAV, WavSettings * AIFF, AiffSettings * AC3, Ac3Settings * EAC3, Eac3Settings * EAC3_ATMOS, Eac3AtmosSettings
*/
CodecSettings?: AudioCodecSettings;
/**
* Specify the language for this audio output track, using the ISO 639-2 or ISO 639-3 three-letter language code. The language specified will be used when 'Follow Input Language Code' is not selected or when 'Follow Input Language Code' is selected but there is no ISO 639 language code specified by the input.
*/
CustomLanguageCode?: __stringMin3Max3PatternAZaZ3;
/**
* Indicates the language of the audio output track. The ISO 639 language specified in the 'Language Code' drop down will be used when 'Follow Input Language Code' is not selected or when 'Follow Input Language Code' is selected but there is no ISO 639 language code specified by the input.
*/
LanguageCode?: LanguageCode;
/**
* Choosing FOLLOW_INPUT will cause the ISO 639 language code of the output to follow the ISO 639 language code of the input. The language specified for languageCode' will be used when USE_CONFIGURED is selected or when FOLLOW_INPUT is selected but there is no ISO 639 language code specified by the input.
*/
LanguageCodeControl?: AudioLanguageCodeControl;
/**
* Advanced audio remixing settings.
*/
RemixSettings?: RemixSettings;
/**
* Specify a label for this output audio stream. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
*/
StreamName?: __stringPatternWS;
}
export type AudioLanguageCodeControl = "FOLLOW_INPUT"|"USE_CONFIGURED"|string;
export type AudioNormalizationAlgorithm = "ITU_BS_1770_1"|"ITU_BS_1770_2"|"ITU_BS_1770_3"|"ITU_BS_1770_4"|string;
export type AudioNormalizationAlgorithmControl = "CORRECT_AUDIO"|"MEASURE_ONLY"|string;
export type AudioNormalizationLoudnessLogging = "LOG"|"DONT_LOG"|string;
export type AudioNormalizationPeakCalculation = "TRUE_PEAK"|"NONE"|string;
export interface AudioNormalizationSettings {
/**
* Choose one of the following audio normalization algorithms: ITU-R BS.1770-1: Ungated loudness. A measurement of ungated average loudness for an entire piece of content, suitable for measurement of short-form content under ATSC recommendation A/85. Supports up to 5.1 audio channels. ITU-R BS.1770-2: Gated loudness. A measurement of gated average loudness compliant with the requirements of EBU-R128. Supports up to 5.1 audio channels. ITU-R BS.1770-3: Modified peak. The same loudness measurement algorithm as 1770-2, with an updated true peak measurement. ITU-R BS.1770-4: Higher channel count. Allows for more audio channels than the other algorithms, including configurations such as 7.1.
*/
Algorithm?: AudioNormalizationAlgorithm;
/**
* When enabled the output audio is corrected using the chosen algorithm. If disabled, the audio will be measured but not adjusted.
*/
AlgorithmControl?: AudioNormalizationAlgorithmControl;
/**
* Content measuring above this level will be corrected to the target level. Content measuring below this level will not be corrected. Gating only applies when not using real_time_correction.
*/
CorrectionGateLevel?: __integerMinNegative70Max0;
/**
* If set to LOG, log each output's audio track loudness to a CSV file.
*/
LoudnessLogging?: AudioNormalizationLoudnessLogging;
/**
* If set to TRUE_PEAK, calculate and log the TruePeak for each output's audio track loudness.
*/
PeakCalculation?: AudioNormalizationPeakCalculation;
/**
* When you use Audio normalization (AudioNormalizationSettings), optionally use this setting to specify a target loudness. If you don't specify a value here, the encoder chooses a value for you, based on the algorithm that you choose for Algorithm (algorithm). If you choose algorithm 1770-1, the encoder will choose -24 LKFS; otherwise, the encoder will choose -23 LKFS.
*/
TargetLkfs?: __doubleMinNegative59Max0;
}
export interface AudioSelector {
/**
* Selects a specific language code from within an audio source, using the ISO 639-2 or ISO 639-3 three-letter language code
*/
CustomLanguageCode?: __stringMin3Max3PatternAZaZ3;
/**
* Enable this setting on one audio selector to set it as the default for the job. The service uses this default for outputs where it can't find the specified input audio. If you don't set a default, those outputs have no audio.
*/
DefaultSelection?: AudioDefaultSelection;
/**
* Specifies audio data from an external file source.
*/
ExternalAudioFileInput?: __stringPatternHttpHttpsS3MM2VVMMPPEEGGMMPP3AAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMAAAACCAAIIFFFFMMPP2AACC3EECC3DDTTSSEE;
/**
* Selects a specific language code from within an audio source.
*/
LanguageCode?: LanguageCode;
/**
* Specifies a time delta in milliseconds to offset the audio from the input video.
*/
Offset?: __integerMinNegative2147483648Max2147483647;
/**
* Selects a specific PID from within an audio source (e.g. 257 selects PID 0x101).
*/
Pids?: __listOf__integerMin1Max2147483647;
/**
* Use this setting for input streams that contain Dolby E, to have the service extract specific program data from the track. To select multiple programs, create multiple selectors with the same Track and different Program numbers. In the console, this setting is visible when you set Selector type to Track. Choose the program number from the dropdown list. If you are sending a JSON file, provide the program ID, which is part of the audio metadata. If your input file has incorrect metadata, you can choose All channels instead of a program number to have the service ignore the program IDs and include all the programs in the track.
*/
ProgramSelection?: __integerMin0Max8;
/**
* Use these settings to reorder the audio channels of one input to match those of another input. This allows you to combine the two files into a single output, one after the other.
*/
RemixSettings?: RemixSettings;
/**
* Specifies the type of the audio selector.
*/
SelectorType?: AudioSelectorType;
/**
* Identify a track from the input audio to include in this selector by entering the track index number. To include several tracks in a single audio selector, specify multiple tracks as follows. Using the console, enter a comma-separated list. For examle, type "1,2,3" to include tracks 1 through 3. Specifying directly in your JSON job file, provide the track numbers in an array. For example, "tracks": [1,2,3].
*/
Tracks?: __listOf__integerMin1Max2147483647;
}
export interface AudioSelectorGroup {
/**
* Name of an Audio Selector within the same input to include in the group. Audio selector names are standardized, based on their order within the input (e.g., "Audio Selector 1"). The audio selector name parameter can be repeated to add any number of audio selectors to the group.
*/
AudioSelectorNames?: __listOf__stringMin1;
}
export type AudioSelectorType = "PID"|"TRACK"|"LANGUAGE_CODE"|string;
export type AudioTypeControl = "FOLLOW_INPUT"|"USE_CONFIGURED"|string;
export interface AvailBlanking {
/**
* Blanking image to be used. Leave empty for solid black. Only bmp and png images are supported.
*/
AvailBlankingImage?: __stringMin14PatternHttpHttpsS3BmpBMPPngPNG;
}
export type BillingTagsSource = "QUEUE"|"PRESET"|"JOB_TEMPLATE"|"JOB"|string;
export interface BurninDestinationSettings {
/**
* If no explicit x_position or y_position is provided, setting alignment to centered will place the captions at the bottom center of the output. Similarly, setting a left alignment will align captions to the bottom left of the output. If x and y positions are given in conjunction with the alignment parameter, the font will be justified (either left or centered) relative to those coordinates. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
Alignment?: BurninSubtitleAlignment;
/**
* Specifies the color of the rectangle behind the captions.
All burn-in and DVB-Sub font settings must match.
*/
BackgroundColor?: BurninSubtitleBackgroundColor;
/**
* Specifies the opacity of the background rectangle. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.
*/
BackgroundOpacity?: __integerMin0Max255;
/**
* Specifies the color of the burned-in captions. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
FontColor?: BurninSubtitleFontColor;
/**
* Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.
All burn-in and DVB-Sub font settings must match.
*/
FontOpacity?: __integerMin0Max255;
/**
* Font resolution in DPI (dots per inch); default is 96 dpi.
All burn-in and DVB-Sub font settings must match.
*/
FontResolution?: __integerMin96Max600;
/**
* Provide the font script, using an ISO 15924 script code, if the LanguageCode is not sufficient for determining the script type. Where LanguageCode or CustomLanguageCode is sufficient, use "AUTOMATIC" or leave unset. This is used to help determine the appropriate font for rendering burn-in captions.
*/
FontScript?: FontScript;
/**
* A positive integer indicates the exact font size in points. Set to 0 for automatic font size selection. All burn-in and DVB-Sub font settings must match.
*/
FontSize?: __integerMin0Max96;
/**
* Specifies font outline color. This option is not valid for source captions that are either 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
OutlineColor?: BurninSubtitleOutlineColor;
/**
* Specifies font outline size in pixels. This option is not valid for source captions that are either 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
OutlineSize?: __integerMin0Max10;
/**
* Specifies the color of the shadow cast by the captions.
All burn-in and DVB-Sub font settings must match.
*/
ShadowColor?: BurninSubtitleShadowColor;
/**
* Specifies the opacity of the shadow. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.
*/
ShadowOpacity?: __integerMin0Max255;
/**
* Specifies the horizontal offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match.
*/
ShadowXOffset?: __integerMinNegative2147483648Max2147483647;
/**
* Specifies the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match.
*/
ShadowYOffset?: __integerMinNegative2147483648Max2147483647;
/**
* Only applies to jobs with input captions in Teletext or STL formats. Specify whether the spacing between letters in your captions is set by the captions grid or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read if the captions are closed caption.
*/
TeletextSpacing?: BurninSubtitleTeletextSpacing;
/**
* Specifies the horizontal position of the caption relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
XPosition?: __integerMin0Max2147483647;
/**
* Specifies the vertical position of the caption relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.
*/
YPosition?: __integerMin0Max2147483647;
}
export type BurninSubtitleAlignment = "CENTERED"|"LEFT"|string;
export type BurninSubtitleBackgroundColor = "NONE"|"BLACK"|"WHITE"|string;
export type BurninSubtitleFontColor = "WHITE"|"BLACK"|"YELLOW"|"RED"|"GREEN"|"BLUE"|string;
export type BurninSubtitleOutlineColor = "BLACK"|"WHITE"|"YELLOW"|"RED"|"GREEN"|"BLUE"|string;
export type BurninSubtitleShadowColor = "NONE"|"BLACK"|"WHITE"|string;
export type BurninSubtitleTeletextSpacing = "FIXED_GRID"|"PROPORTIONAL"|string;
export interface CancelJobRequest {
/**
* The Job ID of the job to be cancelled.
*/
Id: __string;
}
export interface CancelJobResponse {
}
export interface CaptionDescription {
/**
* Specifies which "Caption Selector":#inputs-caption_selector to use from each input when generating captions. The name should be of the format "Caption Selector ", which denotes that the Nth Caption Selector will be used from each input.
*/
CaptionSelectorName?: __stringMin1;
/**
* Indicates the language of the caption output track, using the ISO 639-2 or ISO 639-3 three-letter language code. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
*/
CustomLanguageCode?: __stringMin3Max3PatternAZaZ3;
/**
* Specific settings required by destination type. Note that burnin_destination_settings are not available if the source of the caption data is Embedded or Teletext.
*/
DestinationSettings?: CaptionDestinationSettings;
/**
* Specify the language of this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
*/
LanguageCode?: LanguageCode;
/**
* Specify a label for this set of output captions. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
*/
LanguageDescription?: __string;
}
export interface CaptionDescriptionPreset {
/**
* Indicates the language of the caption output track, using the ISO 639-2 or ISO 639-3 three-letter language code. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
*/
CustomLanguageCode?: __stringMin3Max3PatternAZaZ3;
/**
* Specific settings required by destination type. Note that burnin_destination_settings are not available if the source of the caption data is Embedded or Teletext.
*/
DestinationSettings?: CaptionDestinationSettings;
/**
* Specify the language of this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
*/
LanguageCode?: LanguageCode;
/**
* Specify a label for this set of output captions. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
*/
LanguageDescription?: __string;
}
export interface CaptionDestinationSettings {
/**
* Burn-In Destination Settings.
*/
BurninDestinationSettings?: BurninDestinationSettings;
/**
* Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Other options are embedded with SCTE-20, burn-in, DVB-sub, IMSC, SCC, SRT, teletext, TTML, and web-VTT. If you are using SCTE-20, choose SCTE-20 plus embedded (SCTE20_PLUS_EMBEDDED) to create an output that complies with the SCTE-43 spec. To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20 (EMBEDDED_PLUS_SCTE20).
*/
DestinationType?: CaptionDestinationType;
/**
* DVB-Sub Destination Settings
*/
DvbSubDestinationSettings?: DvbSubDestinationSettings;
/**
* Settings specific to embedded/ancillary caption outputs, including 608/708 Channel destination number.
*/
EmbeddedDestinationSettings?: EmbeddedDestinationSettings;
/**
* Settings specific to IMSC caption outputs.
*/
ImscDestinationSettings?: ImscDestinationSettings;
/**
* Settings for SCC caption output.
*/
SccDestinationSettings?: SccDestinationSettings;
/**
* Settings for Teletext caption output
*/
TeletextDestinationSettings?: TeletextDestinationSettings;
/**
* Settings specific to TTML caption outputs, including Pass style information (TtmlStylePassthrough).
*/
TtmlDestinationSettings?: TtmlDestinationSettings;
}
export type CaptionDestinationType = "BURN_IN"|"DVB_SUB"|"EMBEDDED"|"EMBEDDED_PLUS_SCTE20"|"IMSC"|"SCTE20_PLUS_EMBEDDED"|"SCC"|"SRT"|"SMI"|"TELETEXT"|"TTML"|"WEBVTT"|string;
export interface CaptionSelector {
/**
* The specific language to extract from source, using the ISO 639-2 or ISO 639-3 three-letter language code. If input is SCTE-27, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub and output is Burn-in or SMPTE-TT, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub that is being passed through, omit this field (and PID field); there is no way to extract a specific language with pass-through captions.
*/
CustomLanguageCode?: __stringMin3Max3PatternAZaZ3;
/**
* The specific language to extract from source. If input is SCTE-27, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub and output is Burn-in or SMPTE-TT, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub that is being passed through, omit this field (and PID field); there is no way to extract a specific language with pass-through captions.
*/
LanguageCode?: LanguageCode;
/**
* If your input captions are SCC, TTML, STL, SMI, SRT, or IMSC in an xml file, specify the URI of the input captions source file. If your input captions are IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings.
*/
SourceSettings?: CaptionSourceSettings;
}
export interface CaptionSourceSettings {
/**
* Settings for ancillary captions source.
*/
AncillarySourceSettings?: AncillarySourceSettings;
/**
* DVB Sub Source Settings
*/
DvbSubSourceSettings?: DvbSubSourceSettings;
/**
* Settings for embedded captions Source
*/
EmbeddedSourceSettings?: EmbeddedSourceSettings;
/**
* If your input captions are SCC, SMI, SRT, STL, TTML, or IMSC 1.1 in an xml file, specify the URI of the input caption source file. If your caption source is IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings.
*/
FileSourceSettings?: FileSourceSettings;
/**
* Use Source (SourceType) to identify the format of your input captions. The service cannot auto-detect caption format.
*/
SourceType?: CaptionSourceType;
/**
* Settings specific to Teletext caption sources, including Page number.
*/
TeletextSourceSettings?: TeletextSourceSettings;
/**
* Settings specific to caption sources that are specified by track number. Currently, this is only IMSC captions in an IMF package. If your caption source is IMSC 1.1 in a separate xml file, use FileSourceSettings instead of TrackSourceSettings.
*/
TrackSourceSettings?: TrackSourceSettings;
}
export type CaptionSourceType = "ANCILLARY"|"DVB_SUB"|"EMBEDDED"|"SCTE20"|"SCC"|"TTML"|"STL"|"SRT"|"SMI"|"TELETEXT"|"NULL_SOURCE"|"IMSC"|string;
export interface ChannelMapping {
/**
* List of output channels
*/
OutputChannels?: __listOfOutputChannelMapping;
}
export type CmafClientCache