1. Packages
  2. Azure Native v1
  3. API Docs
  4. media
  5. Transform
These are the docs for Azure Native v1. We recommenend using the latest version, Azure Native v2.
Azure Native v1 v1.104.0 published on Thursday, Jul 6, 2023 by Pulumi

azure-native.media.Transform

Explore with Pulumi AI

azure-native-v1 logo
These are the docs for Azure Native v1. We recommenend using the latest version, Azure Native v2.
Azure Native v1 v1.104.0 published on Thursday, Jul 6, 2023 by Pulumi

    A Transform encapsulates the rules or instructions for generating desired outputs from input media, such as by transcoding or by extracting insights. After the Transform is created, it can be applied to input media by creating Jobs. API Version: 2020-05-01.

    Example Usage

    Create or update a Transform

    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using AzureNative = Pulumi.AzureNative;
    
    return await Deployment.RunAsync(() => 
    {
        var transform = new AzureNative.Media.Transform("transform", new()
        {
            AccountName = "contosomedia",
            Description = "Example Transform to illustrate create and update.",
            Outputs = new[]
            {
                new AzureNative.Media.Inputs.TransformOutputArgs
                {
                    Preset = new AzureNative.Media.Inputs.BuiltInStandardEncoderPresetArgs
                    {
                        OdataType = "#Microsoft.Media.BuiltInStandardEncoderPreset",
                        PresetName = "AdaptiveStreaming",
                    },
                },
            },
            ResourceGroupName = "contosoresources",
            TransformName = "createdTransform",
        });
    
    });
    
    package main
    
    import (
    	media "github.com/pulumi/pulumi-azure-native-sdk/media"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		_, err := media.NewTransform(ctx, "transform", &media.TransformArgs{
    			AccountName: pulumi.String("contosomedia"),
    			Description: pulumi.String("Example Transform to illustrate create and update."),
    			Outputs: []media.TransformOutputTypeArgs{
    				{
    					Preset: {
    						OdataType:  "#Microsoft.Media.BuiltInStandardEncoderPreset",
    						PresetName: "AdaptiveStreaming",
    					},
    				},
    			},
    			ResourceGroupName: pulumi.String("contosoresources"),
    			TransformName:     pulumi.String("createdTransform"),
    		})
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.azurenative.media.Transform;
    import com.pulumi.azurenative.media.TransformArgs;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var transform = new Transform("transform", TransformArgs.builder()        
                .accountName("contosomedia")
                .description("Example Transform to illustrate create and update.")
                .outputs(Map.of("preset", Map.ofEntries(
                    Map.entry("odataType", "#Microsoft.Media.BuiltInStandardEncoderPreset"),
                    Map.entry("presetName", "AdaptiveStreaming")
                )))
                .resourceGroupName("contosoresources")
                .transformName("createdTransform")
                .build());
    
        }
    }
    
    import pulumi
    import pulumi_azure_native as azure_native
    
    transform = azure_native.media.Transform("transform",
        account_name="contosomedia",
        description="Example Transform to illustrate create and update.",
        outputs=[azure_native.media.TransformOutputArgs(
            preset=azure_native.media.BuiltInStandardEncoderPresetArgs(
                odata_type="#Microsoft.Media.BuiltInStandardEncoderPreset",
                preset_name="AdaptiveStreaming",
            ),
        )],
        resource_group_name="contosoresources",
        transform_name="createdTransform")
    
    import * as pulumi from "@pulumi/pulumi";
    import * as azure_native from "@pulumi/azure-native";
    
    const transform = new azure_native.media.Transform("transform", {
        accountName: "contosomedia",
        description: "Example Transform to illustrate create and update.",
        outputs: [{
            preset: {
                odataType: "#Microsoft.Media.BuiltInStandardEncoderPreset",
                presetName: "AdaptiveStreaming",
            },
        }],
        resourceGroupName: "contosoresources",
        transformName: "createdTransform",
    });
    
    resources:
      transform:
        type: azure-native:media:Transform
        properties:
          accountName: contosomedia
          description: Example Transform to illustrate create and update.
          outputs:
            - preset:
                odataType: '#Microsoft.Media.BuiltInStandardEncoderPreset'
                presetName: AdaptiveStreaming
          resourceGroupName: contosoresources
          transformName: createdTransform
    

    Create Transform Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new Transform(name: string, args: TransformArgs, opts?: CustomResourceOptions);
    @overload
    def Transform(resource_name: str,
                  args: TransformArgs,
                  opts: Optional[ResourceOptions] = None)
    
    @overload
    def Transform(resource_name: str,
                  opts: Optional[ResourceOptions] = None,
                  account_name: Optional[str] = None,
                  outputs: Optional[Sequence[TransformOutputArgs]] = None,
                  resource_group_name: Optional[str] = None,
                  description: Optional[str] = None,
                  transform_name: Optional[str] = None)
    func NewTransform(ctx *Context, name string, args TransformArgs, opts ...ResourceOption) (*Transform, error)
    public Transform(string name, TransformArgs args, CustomResourceOptions? opts = null)
    public Transform(String name, TransformArgs args)
    public Transform(String name, TransformArgs args, CustomResourceOptions options)
    
    type: azure-native:media:Transform
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args TransformArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args TransformArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args TransformArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args TransformArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args TransformArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var transformResource = new AzureNative.Media.Transform("transformResource", new()
    {
        AccountName = "string",
        Outputs = new[]
        {
            
            {
                { "preset", 
                {
                    { "odataType", "#Microsoft.Media.AudioAnalyzerPreset" },
                    { "audioLanguage", "string" },
                    { "experimentalOptions", 
                    {
                        { "string", "string" },
                    } },
                    { "mode", "string" },
                } },
                { "onError", "string" },
                { "relativePriority", "string" },
            },
        },
        ResourceGroupName = "string",
        Description = "string",
        TransformName = "string",
    });
    
    example, err := media.NewTransform(ctx, "transformResource", &media.TransformArgs{
    	AccountName: "string",
    	Outputs: []map[string]interface{}{
    		map[string]interface{}{
    			"preset": map[string]interface{}{
    				"odataType":     "#Microsoft.Media.AudioAnalyzerPreset",
    				"audioLanguage": "string",
    				"experimentalOptions": map[string]interface{}{
    					"string": "string",
    				},
    				"mode": "string",
    			},
    			"onError":          "string",
    			"relativePriority": "string",
    		},
    	},
    	ResourceGroupName: "string",
    	Description:       "string",
    	TransformName:     "string",
    })
    
    var transformResource = new Transform("transformResource", TransformArgs.builder()
        .accountName("string")
        .outputs(%!v(PANIC=Format method: runtime error: invalid memory address or nil pointer dereference))
        .resourceGroupName("string")
        .description("string")
        .transformName("string")
        .build());
    
    transform_resource = azure_native.media.Transform("transformResource",
        account_name=string,
        outputs=[{
            preset: {
                odataType: #Microsoft.Media.AudioAnalyzerPreset,
                audioLanguage: string,
                experimentalOptions: {
                    string: string,
                },
                mode: string,
            },
            onError: string,
            relativePriority: string,
        }],
        resource_group_name=string,
        description=string,
        transform_name=string)
    
    const transformResource = new azure_native.media.Transform("transformResource", {
        accountName: "string",
        outputs: [{
            preset: {
                odataType: "#Microsoft.Media.AudioAnalyzerPreset",
                audioLanguage: "string",
                experimentalOptions: {
                    string: "string",
                },
                mode: "string",
            },
            onError: "string",
            relativePriority: "string",
        }],
        resourceGroupName: "string",
        description: "string",
        transformName: "string",
    });
    
    type: azure-native:media:Transform
    properties:
        accountName: string
        description: string
        outputs:
            - onError: string
              preset:
                audioLanguage: string
                experimentalOptions:
                    string: string
                mode: string
                odataType: '#Microsoft.Media.AudioAnalyzerPreset'
              relativePriority: string
        resourceGroupName: string
        transformName: string
    

    Transform Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    The Transform resource accepts the following input properties:

    AccountName string
    The Media Services account name.
    Outputs List<Pulumi.AzureNative.Media.Inputs.TransformOutput>
    An array of one or more TransformOutputs that the Transform should generate.
    ResourceGroupName string
    The name of the resource group within the Azure subscription.
    Description string
    An optional verbose description of the Transform.
    TransformName string
    The Transform name.
    AccountName string
    The Media Services account name.
    Outputs []TransformOutputTypeArgs
    An array of one or more TransformOutputs that the Transform should generate.
    ResourceGroupName string
    The name of the resource group within the Azure subscription.
    Description string
    An optional verbose description of the Transform.
    TransformName string
    The Transform name.
    accountName String
    The Media Services account name.
    outputs List<TransformOutput>
    An array of one or more TransformOutputs that the Transform should generate.
    resourceGroupName String
    The name of the resource group within the Azure subscription.
    description String
    An optional verbose description of the Transform.
    transformName String
    The Transform name.
    accountName string
    The Media Services account name.
    outputs TransformOutput[]
    An array of one or more TransformOutputs that the Transform should generate.
    resourceGroupName string
    The name of the resource group within the Azure subscription.
    description string
    An optional verbose description of the Transform.
    transformName string
    The Transform name.
    account_name str
    The Media Services account name.
    outputs Sequence[TransformOutputArgs]
    An array of one or more TransformOutputs that the Transform should generate.
    resource_group_name str
    The name of the resource group within the Azure subscription.
    description str
    An optional verbose description of the Transform.
    transform_name str
    The Transform name.
    accountName String
    The Media Services account name.
    outputs List<Property Map>
    An array of one or more TransformOutputs that the Transform should generate.
    resourceGroupName String
    The name of the resource group within the Azure subscription.
    description String
    An optional verbose description of the Transform.
    transformName String
    The Transform name.

    Outputs

    All input properties are implicitly available as output properties. Additionally, the Transform resource produces the following output properties:

    Created string
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    Id string
    The provider-assigned unique ID for this managed resource.
    LastModified string
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    Name string
    The name of the resource
    SystemData Pulumi.AzureNative.Media.Outputs.SystemDataResponse
    The system metadata relating to this resource.
    Type string
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
    Created string
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    Id string
    The provider-assigned unique ID for this managed resource.
    LastModified string
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    Name string
    The name of the resource
    SystemData SystemDataResponse
    The system metadata relating to this resource.
    Type string
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
    created String
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    id String
    The provider-assigned unique ID for this managed resource.
    lastModified String
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    name String
    The name of the resource
    systemData SystemDataResponse
    The system metadata relating to this resource.
    type String
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
    created string
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    id string
    The provider-assigned unique ID for this managed resource.
    lastModified string
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    name string
    The name of the resource
    systemData SystemDataResponse
    The system metadata relating to this resource.
    type string
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
    created str
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    id str
    The provider-assigned unique ID for this managed resource.
    last_modified str
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    name str
    The name of the resource
    system_data SystemDataResponse
    The system metadata relating to this resource.
    type str
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
    created String
    The UTC date and time when the Transform was created, in 'YYYY-MM-DDThh:mm:ssZ' format.
    id String
    The provider-assigned unique ID for this managed resource.
    lastModified String
    The UTC date and time when the Transform was last updated, in 'YYYY-MM-DDThh:mm:ssZ' format.
    name String
    The name of the resource
    systemData Property Map
    The system metadata relating to this resource.
    type String
    The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"

    Supporting Types

    AacAudio, AacAudioArgs

    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Profile string | Pulumi.AzureNative.Media.AacAudioProfile
    The encoding profile to be used when encoding audio with AAC.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Profile string | AacAudioProfile
    The encoding profile to be used when encoding audio with AAC.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    bitrate Integer
    The bitrate, in bits per second, of the output encoded audio.
    channels Integer
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    profile String | AacAudioProfile
    The encoding profile to be used when encoding audio with AAC.
    samplingRate Integer
    The sampling rate to use for encoding in hertz.
    bitrate number
    The bitrate, in bits per second, of the output encoded audio.
    channels number
    The number of channels in the audio.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    profile string | AacAudioProfile
    The encoding profile to be used when encoding audio with AAC.
    samplingRate number
    The sampling rate to use for encoding in hertz.
    bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    channels int
    The number of channels in the audio.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    profile str | AacAudioProfile
    The encoding profile to be used when encoding audio with AAC.
    sampling_rate int
    The sampling rate to use for encoding in hertz.
    bitrate Number
    The bitrate, in bits per second, of the output encoded audio.
    channels Number
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    profile String | "AacLc" | "HeAacV1" | "HeAacV2"
    The encoding profile to be used when encoding audio with AAC.
    samplingRate Number
    The sampling rate to use for encoding in hertz.

    AacAudioProfile, AacAudioProfileArgs

    AacLc
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    HeAacV1
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    HeAacV2
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
    AacAudioProfileAacLc
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    AacAudioProfileHeAacV1
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    AacAudioProfileHeAacV2
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
    AacLc
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    HeAacV1
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    HeAacV2
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
    AacLc
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    HeAacV1
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    HeAacV2
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
    AAC_LC
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    HE_AAC_V1
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    HE_AAC_V2
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.
    "AacLc"
    AacLcSpecifies that the output audio is to be encoded into AAC Low Complexity profile (AAC-LC).
    "HeAacV1"
    HeAacV1Specifies that the output audio is to be encoded into HE-AAC v1 profile.
    "HeAacV2"
    HeAacV2Specifies that the output audio is to be encoded into HE-AAC v2 profile.

    AacAudioResponse, AacAudioResponseArgs

    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Profile string
    The encoding profile to be used when encoding audio with AAC.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Profile string
    The encoding profile to be used when encoding audio with AAC.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    bitrate Integer
    The bitrate, in bits per second, of the output encoded audio.
    channels Integer
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    profile String
    The encoding profile to be used when encoding audio with AAC.
    samplingRate Integer
    The sampling rate to use for encoding in hertz.
    bitrate number
    The bitrate, in bits per second, of the output encoded audio.
    channels number
    The number of channels in the audio.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    profile string
    The encoding profile to be used when encoding audio with AAC.
    samplingRate number
    The sampling rate to use for encoding in hertz.
    bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    channels int
    The number of channels in the audio.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    profile str
    The encoding profile to be used when encoding audio with AAC.
    sampling_rate int
    The sampling rate to use for encoding in hertz.
    bitrate Number
    The bitrate, in bits per second, of the output encoded audio.
    channels Number
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    profile String
    The encoding profile to be used when encoding audio with AAC.
    samplingRate Number
    The sampling rate to use for encoding in hertz.

    AnalysisResolution, AnalysisResolutionArgs

    SourceResolution
    SourceResolution
    StandardDefinition
    StandardDefinition
    AnalysisResolutionSourceResolution
    SourceResolution
    AnalysisResolutionStandardDefinition
    StandardDefinition
    SourceResolution
    SourceResolution
    StandardDefinition
    StandardDefinition
    SourceResolution
    SourceResolution
    StandardDefinition
    StandardDefinition
    SOURCE_RESOLUTION
    SourceResolution
    STANDARD_DEFINITION
    StandardDefinition
    "SourceResolution"
    SourceResolution
    "StandardDefinition"
    StandardDefinition

    Audio, AudioArgs

    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    bitrate Integer
    The bitrate, in bits per second, of the output encoded audio.
    channels Integer
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate Integer
    The sampling rate to use for encoding in hertz.
    bitrate number
    The bitrate, in bits per second, of the output encoded audio.
    channels number
    The number of channels in the audio.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate number
    The sampling rate to use for encoding in hertz.
    bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    channels int
    The number of channels in the audio.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    sampling_rate int
    The sampling rate to use for encoding in hertz.
    bitrate Number
    The bitrate, in bits per second, of the output encoded audio.
    channels Number
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate Number
    The sampling rate to use for encoding in hertz.

    AudioAnalysisMode, AudioAnalysisModeArgs

    Standard
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    Basic
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
    AudioAnalysisModeStandard
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    AudioAnalysisModeBasic
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
    Standard
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    Basic
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
    Standard
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    Basic
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
    STANDARD
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    BASIC
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.
    "Standard"
    StandardPerforms all operations included in the Basic mode, additionally performing language detection and speaker diarization.
    "Basic"
    BasicThis mode performs speech-to-text transcription and generation of a VTT subtitle/caption file. The output of this mode includes an Insights JSON file including only the keywords, transcription,and timing information. Automatic language detection and speaker diarization are not included in this mode.

    AudioAnalyzerPreset, AudioAnalyzerPresetArgs

    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string | Pulumi.AzureNative.Media.AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode string | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audio_language str
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode str | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String | "Standard" | "Basic"
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

    AudioAnalyzerPresetResponse, AudioAnalyzerPresetResponseArgs

    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audio_language str
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode str
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

    AudioOverlay, AudioOverlayArgs

    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel float64
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    input_label str
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audio_gain_level float
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end str
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fade_in_duration str
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fade_out_duration str
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start str
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

    AudioOverlayResponse, AudioOverlayResponseArgs

    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel float64
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    input_label str
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audio_gain_level float
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end str
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fade_in_duration str
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fade_out_duration str
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start str
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

    AudioResponse, AudioResponseArgs

    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    Bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    Channels int
    The number of channels in the audio.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    SamplingRate int
    The sampling rate to use for encoding in hertz.
    bitrate Integer
    The bitrate, in bits per second, of the output encoded audio.
    channels Integer
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate Integer
    The sampling rate to use for encoding in hertz.
    bitrate number
    The bitrate, in bits per second, of the output encoded audio.
    channels number
    The number of channels in the audio.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate number
    The sampling rate to use for encoding in hertz.
    bitrate int
    The bitrate, in bits per second, of the output encoded audio.
    channels int
    The number of channels in the audio.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    sampling_rate int
    The sampling rate to use for encoding in hertz.
    bitrate Number
    The bitrate, in bits per second, of the output encoded audio.
    channels Number
    The number of channels in the audio.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    samplingRate Number
    The sampling rate to use for encoding in hertz.

    BlurType, BlurTypeArgs

    Box
    BoxBox: debug filter, bounding box only
    Low
    LowLow: box-car blur filter
    Med
    MedMed: Gaussian blur filter
    High
    HighHigh: Confuse blur filter
    Black
    BlackBlack: Black out filter
    BlurTypeBox
    BoxBox: debug filter, bounding box only
    BlurTypeLow
    LowLow: box-car blur filter
    BlurTypeMed
    MedMed: Gaussian blur filter
    BlurTypeHigh
    HighHigh: Confuse blur filter
    BlurTypeBlack
    BlackBlack: Black out filter
    Box
    BoxBox: debug filter, bounding box only
    Low
    LowLow: box-car blur filter
    Med
    MedMed: Gaussian blur filter
    High
    HighHigh: Confuse blur filter
    Black
    BlackBlack: Black out filter
    Box
    BoxBox: debug filter, bounding box only
    Low
    LowLow: box-car blur filter
    Med
    MedMed: Gaussian blur filter
    High
    HighHigh: Confuse blur filter
    Black
    BlackBlack: Black out filter
    BOX
    BoxBox: debug filter, bounding box only
    LOW
    LowLow: box-car blur filter
    MED
    MedMed: Gaussian blur filter
    HIGH
    HighHigh: Confuse blur filter
    BLACK
    BlackBlack: Black out filter
    "Box"
    BoxBox: debug filter, bounding box only
    "Low"
    LowLow: box-car blur filter
    "Med"
    MedMed: Gaussian blur filter
    "High"
    HighHigh: Confuse blur filter
    "Black"
    BlackBlack: Black out filter

    BuiltInStandardEncoderPreset, BuiltInStandardEncoderPresetArgs

    PresetName string | Pulumi.AzureNative.Media.EncoderNamedPreset
    The built-in preset to be used for encoding videos.
    PresetName string | EncoderNamedPreset
    The built-in preset to be used for encoding videos.
    presetName String | EncoderNamedPreset
    The built-in preset to be used for encoding videos.
    presetName string | EncoderNamedPreset
    The built-in preset to be used for encoding videos.
    preset_name str | EncoderNamedPreset
    The built-in preset to be used for encoding videos.

    BuiltInStandardEncoderPresetResponse, BuiltInStandardEncoderPresetResponseArgs

    PresetName string
    The built-in preset to be used for encoding videos.
    PresetName string
    The built-in preset to be used for encoding videos.
    presetName String
    The built-in preset to be used for encoding videos.
    presetName string
    The built-in preset to be used for encoding videos.
    preset_name str
    The built-in preset to be used for encoding videos.
    presetName String
    The built-in preset to be used for encoding videos.

    CopyAudio, CopyAudioArgs

    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.

    CopyAudioResponse, CopyAudioResponseArgs

    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.

    CopyVideo, CopyVideoArgs

    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.

    CopyVideoResponse, CopyVideoResponseArgs

    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.

    Deinterlace, DeinterlaceArgs

    Mode string | Pulumi.AzureNative.Media.DeinterlaceMode
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    Parity string | Pulumi.AzureNative.Media.DeinterlaceParity
    The field parity for de-interlacing, defaults to Auto.
    Mode string | DeinterlaceMode
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    Parity string | DeinterlaceParity
    The field parity for de-interlacing, defaults to Auto.
    mode String | DeinterlaceMode
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity String | DeinterlaceParity
    The field parity for de-interlacing, defaults to Auto.
    mode string | DeinterlaceMode
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity string | DeinterlaceParity
    The field parity for de-interlacing, defaults to Auto.
    mode str | DeinterlaceMode
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity str | DeinterlaceParity
    The field parity for de-interlacing, defaults to Auto.
    mode String | "Off" | "AutoPixelAdaptive"
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity String | "Auto" | "TopFieldFirst" | "BottomFieldFirst"
    The field parity for de-interlacing, defaults to Auto.

    DeinterlaceMode, DeinterlaceModeArgs

    Off
    OffDisables de-interlacing of the source video.
    AutoPixelAdaptive
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
    DeinterlaceModeOff
    OffDisables de-interlacing of the source video.
    DeinterlaceModeAutoPixelAdaptive
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
    Off
    OffDisables de-interlacing of the source video.
    AutoPixelAdaptive
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
    Off
    OffDisables de-interlacing of the source video.
    AutoPixelAdaptive
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
    OFF
    OffDisables de-interlacing of the source video.
    AUTO_PIXEL_ADAPTIVE
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.
    "Off"
    OffDisables de-interlacing of the source video.
    "AutoPixelAdaptive"
    AutoPixelAdaptiveApply automatic pixel adaptive de-interlacing on each frame in the input video.

    DeinterlaceParity, DeinterlaceParityArgs

    Auto
    AutoAutomatically detect the order of fields
    TopFieldFirst
    TopFieldFirstApply top field first processing of input video.
    BottomFieldFirst
    BottomFieldFirstApply bottom field first processing of input video.
    DeinterlaceParityAuto
    AutoAutomatically detect the order of fields
    DeinterlaceParityTopFieldFirst
    TopFieldFirstApply top field first processing of input video.
    DeinterlaceParityBottomFieldFirst
    BottomFieldFirstApply bottom field first processing of input video.
    Auto
    AutoAutomatically detect the order of fields
    TopFieldFirst
    TopFieldFirstApply top field first processing of input video.
    BottomFieldFirst
    BottomFieldFirstApply bottom field first processing of input video.
    Auto
    AutoAutomatically detect the order of fields
    TopFieldFirst
    TopFieldFirstApply top field first processing of input video.
    BottomFieldFirst
    BottomFieldFirstApply bottom field first processing of input video.
    AUTO
    AutoAutomatically detect the order of fields
    TOP_FIELD_FIRST
    TopFieldFirstApply top field first processing of input video.
    BOTTOM_FIELD_FIRST
    BottomFieldFirstApply bottom field first processing of input video.
    "Auto"
    AutoAutomatically detect the order of fields
    "TopFieldFirst"
    TopFieldFirstApply top field first processing of input video.
    "BottomFieldFirst"
    BottomFieldFirstApply bottom field first processing of input video.

    DeinterlaceResponse, DeinterlaceResponseArgs

    Mode string
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    Parity string
    The field parity for de-interlacing, defaults to Auto.
    Mode string
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    Parity string
    The field parity for de-interlacing, defaults to Auto.
    mode String
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity String
    The field parity for de-interlacing, defaults to Auto.
    mode string
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity string
    The field parity for de-interlacing, defaults to Auto.
    mode str
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity str
    The field parity for de-interlacing, defaults to Auto.
    mode String
    The deinterlacing mode. Defaults to AutoPixelAdaptive.
    parity String
    The field parity for de-interlacing, defaults to Auto.

    EncoderNamedPreset, EncoderNamedPresetArgs

    H264SingleBitrateSD
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate720p
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate1080p
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    AdaptiveStreaming
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    AACGoodQualityAudio
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    ContentAwareEncodingExperimental
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    ContentAwareEncoding
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    CopyAllBitrateNonInterleaved
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    H264MultipleBitrate1080p
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    H264MultipleBitrate720p
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    H264MultipleBitrateSD
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    H265ContentAwareEncoding
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    H265AdaptiveStreaming
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    H265SingleBitrate720p
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate1080p
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate4K
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetH264SingleBitrateSD
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetH264SingleBitrate720p
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetH264SingleBitrate1080p
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetAdaptiveStreaming
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    EncoderNamedPresetAACGoodQualityAudio
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    EncoderNamedPresetContentAwareEncodingExperimental
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    EncoderNamedPresetContentAwareEncoding
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    EncoderNamedPresetCopyAllBitrateNonInterleaved
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    EncoderNamedPresetH264MultipleBitrate1080p
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    EncoderNamedPresetH264MultipleBitrate720p
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    EncoderNamedPresetH264MultipleBitrateSD
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    EncoderNamedPresetH265ContentAwareEncoding
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    EncoderNamedPresetH265AdaptiveStreaming
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    EncoderNamedPresetH265SingleBitrate720p
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetH265SingleBitrate1080p
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    EncoderNamedPresetH265SingleBitrate4K
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrateSD
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate720p
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate1080p
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    AdaptiveStreaming
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    AACGoodQualityAudio
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    ContentAwareEncodingExperimental
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    ContentAwareEncoding
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    CopyAllBitrateNonInterleaved
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    H264MultipleBitrate1080p
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    H264MultipleBitrate720p
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    H264MultipleBitrateSD
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    H265ContentAwareEncoding
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    H265AdaptiveStreaming
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    H265SingleBitrate720p
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate1080p
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate4K
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrateSD
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate720p
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264SingleBitrate1080p
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    AdaptiveStreaming
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    AACGoodQualityAudio
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    ContentAwareEncodingExperimental
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    ContentAwareEncoding
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    CopyAllBitrateNonInterleaved
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    H264MultipleBitrate1080p
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    H264MultipleBitrate720p
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    H264MultipleBitrateSD
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    H265ContentAwareEncoding
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    H265AdaptiveStreaming
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    H265SingleBitrate720p
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate1080p
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265SingleBitrate4K
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264_SINGLE_BITRATE_SD
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264_SINGLE_BITRATE720P
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H264_SINGLE_BITRATE1080P
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    ADAPTIVE_STREAMING
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    AAC_GOOD_QUALITY_AUDIO
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    CONTENT_AWARE_ENCODING_EXPERIMENTAL
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    CONTENT_AWARE_ENCODING
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    COPY_ALL_BITRATE_NON_INTERLEAVED
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    H264_MULTIPLE_BITRATE1080P
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    H264_MULTIPLE_BITRATE720P
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    H264_MULTIPLE_BITRATE_SD
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    H265_CONTENT_AWARE_ENCODING
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    H265_ADAPTIVE_STREAMING
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    H265_SINGLE_BITRATE720P
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265_SINGLE_BITRATE1080P
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    H265_SINGLE_BITRATE4_K
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "H264SingleBitrateSD"
    H264SingleBitrateSDProduces an MP4 file where the video is encoded with H.264 codec at 2200 kbps and a picture height of 480 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "H264SingleBitrate720p"
    H264SingleBitrate720pProduces an MP4 file where the video is encoded with H.264 codec at 4500 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "H264SingleBitrate1080p"
    H264SingleBitrate1080pProduces an MP4 file where the video is encoded with H.264 codec at 6750 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "AdaptiveStreaming"
    AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.264 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    "AACGoodQualityAudio"
    AACGoodQualityAudioProduces a single MP4 file containing only stereo audio encoded at 192 kbps.
    "ContentAwareEncodingExperimental"
    ContentAwareEncodingExperimentalExposes an experimental preset for content-aware encoding. Given any input content, the service attempts to automatically determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. The underlying algorithms will continue to evolve over time. The output will contain MP4 files with video and audio interleaved.
    "ContentAwareEncoding"
    ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    "CopyAllBitrateNonInterleaved"
    CopyAllBitrateNonInterleavedCopy all video and audio streams from the input asset as non-interleaved video and audio output files. This preset can be used to clip an existing asset or convert a group of key frame (GOP) aligned MP4 files as an asset that can be streamed.
    "H264MultipleBitrate1080p"
    H264MultipleBitrate1080pProduces a set of 8 GOP-aligned MP4 files, ranging from 6000 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 1080p and goes down to 180p.
    "H264MultipleBitrate720p"
    H264MultipleBitrate720pProduces a set of 6 GOP-aligned MP4 files, ranging from 3400 kbps to 400 kbps, and stereo AAC audio. Resolution starts at 720p and goes down to 180p.
    "H264MultipleBitrateSD"
    H264MultipleBitrateSDProduces a set of 5 GOP-aligned MP4 files, ranging from 1900kbps to 400 kbps, and stereo AAC audio. Resolution starts at 480p and goes down to 240p.
    "H265ContentAwareEncoding"
    H265ContentAwareEncodingProduces a set of GOP-aligned MP4s by using content-aware encoding. Given any input content, the service performs an initial lightweight analysis of the input content, and uses the results to determine the optimal number of layers, appropriate bitrate and resolution settings for delivery by adaptive streaming. This preset is particularly effective for low and medium complexity videos, where the output files will be at lower bitrates but at a quality that still delivers a good experience to viewers. The output will contain MP4 files with video and audio interleaved.
    "H265AdaptiveStreaming"
    H265AdaptiveStreamingProduces a set of GOP aligned MP4 files with H.265 video and stereo AAC audio. Auto-generates a bitrate ladder based on the input resolution, bitrate and frame rate. The auto-generated preset will never exceed the input resolution. For example, if the input is 720p, output will remain 720p at best.
    "H265SingleBitrate720p"
    H265SingleBitrate720pProduces an MP4 file where the video is encoded with H.265 codec at 1800 kbps and a picture height of 720 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "H265SingleBitrate1080p"
    H265SingleBitrate1080pProduces an MP4 file where the video is encoded with H.265 codec at 3500 kbps and a picture height of 1080 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.
    "H265SingleBitrate4K"
    H265SingleBitrate4KProduces an MP4 file where the video is encoded with H.265 codec at 9500 kbps and a picture height of 2160 pixels, and the stereo audio is encoded with AAC-LC codec at 128 kbps.

    EntropyMode, EntropyModeArgs

    Cabac
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    Cavlc
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
    EntropyModeCabac
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    EntropyModeCavlc
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
    Cabac
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    Cavlc
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
    Cabac
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    Cavlc
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
    CABAC
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    CAVLC
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.
    "Cabac"
    CabacContext Adaptive Binary Arithmetic Coder (CABAC) entropy encoding.
    "Cavlc"
    CavlcContext Adaptive Variable Length Coder (CAVLC) entropy encoding.

    FaceDetectorPreset, FaceDetectorPresetArgs

    BlurType string | Pulumi.AzureNative.Media.BlurType
    Blur type
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string | Pulumi.AzureNative.Media.FaceRedactorMode
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    Resolution string | Pulumi.AzureNative.Media.AnalysisResolution
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    BlurType string | BlurType
    Blur type
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string | FaceRedactorMode
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    Resolution string | AnalysisResolution
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType String | BlurType
    Blur type
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String | FaceRedactorMode
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution String | AnalysisResolution
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType string | BlurType
    Blur type
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode string | FaceRedactorMode
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution string | AnalysisResolution
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blur_type str | BlurType
    Blur type
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode str | FaceRedactorMode
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution str | AnalysisResolution
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType String | "Box" | "Low" | "Med" | "High" | "Black"
    Blur type
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String | "Analyze" | "Redact" | "Combined"
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution String | "SourceResolution" | "StandardDefinition"
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

    FaceDetectorPresetResponse, FaceDetectorPresetResponseArgs

    BlurType string
    Blur type
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    Resolution string
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    BlurType string
    Blur type
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    Mode string
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    Resolution string
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType String
    Blur type
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution String
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType string
    Blur type
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode string
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution string
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blur_type str
    Blur type
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode str
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution str
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.
    blurType String
    Blur type
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    mode String
    This mode provides the ability to choose between the following settings: 1) Analyze - For detection only.This mode generates a metadata JSON file marking appearances of faces throughout the video.Where possible, appearances of the same person are assigned the same ID. 2) Combined - Additionally redacts(blurs) detected faces. 3) Redact - This enables a 2-pass process, allowing for selective redaction of a subset of detected faces.It takes in the metadata file from a prior analyze pass, along with the source video, and a user-selected subset of IDs that require redaction.
    resolution String
    Specifies the maximum resolution at which your video is analyzed. The default behavior is "SourceResolution," which will keep the input video at its original resolution when analyzed. Using "StandardDefinition" will resize input videos to standard definition while preserving the appropriate aspect ratio. It will only resize if the video is of higher resolution. For example, a 1920x1080 input would be scaled to 640x360 before processing. Switching to "StandardDefinition" will reduce the time it takes to process high resolution video. It may also reduce the cost of using this component (see https://azure.microsoft.com/en-us/pricing/details/media-services/#analytics for details). However, faces that end up being too small in the resized video may not be detected.

    FaceRedactorMode, FaceRedactorModeArgs

    Analyze
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    Redact
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    Combined
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
    FaceRedactorModeAnalyze
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    FaceRedactorModeRedact
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    FaceRedactorModeCombined
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
    Analyze
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    Redact
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    Combined
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
    Analyze
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    Redact
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    Combined
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
    ANALYZE
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    REDACT
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    COMBINED
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.
    "Analyze"
    AnalyzeAnalyze mode detects faces and outputs a metadata file with the results. Allows editing of the metadata file before faces are blurred with Redact mode.
    "Redact"
    RedactRedact mode consumes the metadata file from Analyze mode and redacts the faces found.
    "Combined"
    CombinedCombined mode does the Analyze and Redact steps in one pass when editing the analyzed faces is not desired.

    Filters, FiltersArgs

    Crop Pulumi.AzureNative.Media.Inputs.Rectangle
    The parameters for the rectangular window with which to crop the input video.
    Deinterlace Pulumi.AzureNative.Media.Inputs.Deinterlace
    The de-interlacing settings.
    Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlay, Pulumi.AzureNative.Media.Inputs.VideoOverlay>>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    Rotation string | Pulumi.AzureNative.Media.Rotation
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    Crop Rectangle
    The parameters for the rectangular window with which to crop the input video.
    Deinterlace Deinterlace
    The de-interlacing settings.
    Overlays []interface{}
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    Rotation string | Rotation
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop Rectangle
    The parameters for the rectangular window with which to crop the input video.
    deinterlace Deinterlace
    The de-interlacing settings.
    overlays List<Either<AudioOverlay,VideoOverlay>>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation String | Rotation
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop Rectangle
    The parameters for the rectangular window with which to crop the input video.
    deinterlace Deinterlace
    The de-interlacing settings.
    overlays (AudioOverlay | VideoOverlay)[]
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation string | Rotation
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop Rectangle
    The parameters for the rectangular window with which to crop the input video.
    deinterlace Deinterlace
    The de-interlacing settings.
    overlays Sequence[Union[AudioOverlay, VideoOverlay]]
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation str | Rotation
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop Property Map
    The parameters for the rectangular window with which to crop the input video.
    deinterlace Property Map
    The de-interlacing settings.
    overlays List<Property Map | Property Map>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation String | "Auto" | "None" | "Rotate0" | "Rotate90" | "Rotate180" | "Rotate270"
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

    FiltersResponse, FiltersResponseArgs

    Crop Pulumi.AzureNative.Media.Inputs.RectangleResponse
    The parameters for the rectangular window with which to crop the input video.
    Deinterlace Pulumi.AzureNative.Media.Inputs.DeinterlaceResponse
    The de-interlacing settings.
    Overlays List<Union<Pulumi.AzureNative.Media.Inputs.AudioOverlayResponse, Pulumi.AzureNative.Media.Inputs.VideoOverlayResponse>>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    Rotation string
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    Crop RectangleResponse
    The parameters for the rectangular window with which to crop the input video.
    Deinterlace DeinterlaceResponse
    The de-interlacing settings.
    Overlays []interface{}
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    Rotation string
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop RectangleResponse
    The parameters for the rectangular window with which to crop the input video.
    deinterlace DeinterlaceResponse
    The de-interlacing settings.
    overlays List<Either<AudioOverlayResponse,VideoOverlayResponse>>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation String
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop RectangleResponse
    The parameters for the rectangular window with which to crop the input video.
    deinterlace DeinterlaceResponse
    The de-interlacing settings.
    overlays (AudioOverlayResponse | VideoOverlayResponse)[]
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation string
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop RectangleResponse
    The parameters for the rectangular window with which to crop the input video.
    deinterlace DeinterlaceResponse
    The de-interlacing settings.
    overlays Sequence[Union[AudioOverlayResponse, VideoOverlayResponse]]
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation str
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto
    crop Property Map
    The parameters for the rectangular window with which to crop the input video.
    deinterlace Property Map
    The de-interlacing settings.
    overlays List<Property Map | Property Map>
    The properties of overlays to be applied to the input video. These could be audio, image or video overlays.
    rotation String
    The rotation, if any, to be applied to the input video, before it is encoded. Default is Auto

    H264Complexity, H264ComplexityArgs

    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    H264ComplexitySpeed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    H264ComplexityBalanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    H264ComplexityQuality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    SPEED
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    BALANCED
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    QUALITY
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    "Speed"
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    "Balanced"
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    "Quality"
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

    H264Layer, H264LayerArgs

    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    AdaptiveBFrame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    EntropyMode string | Pulumi.AzureNative.Media.EntropyMode
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string | Pulumi.AzureNative.Media.H264VideoProfile
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    AdaptiveBFrame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    EntropyMode string | EntropyMode
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string | H264VideoProfile
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Integer
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame Boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Integer
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode String | EntropyMode
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Integer
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String | H264VideoProfile
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames Integer
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Integer
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate number
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode string | EntropyMode
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile string | H264VideoProfile
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptive_b_frame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    b_frames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    buffer_window str
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropy_mode str | EntropyMode
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frame_rate str
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level str
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    max_bitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile str | H264VideoProfile
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    reference_frames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Number
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame Boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode String | "Cabac" | "Cavlc"
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String | "Auto" | "Baseline" | "Main" | "High" | "High422" | "High444"
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames Number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    H264LayerResponse, H264LayerResponseArgs

    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    AdaptiveBFrame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    EntropyMode string
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    AdaptiveBFrame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    EntropyMode string
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Integer
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame Boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Integer
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode String
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Integer
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames Integer
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Integer
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate number
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode string
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile string
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptive_b_frame bool
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    b_frames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    buffer_window str
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropy_mode str
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frame_rate str
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level str
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    max_bitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile str
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    reference_frames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Number
    The average bitrate in bits per second at which to encode the input video when generating this layer. This is a required field.
    adaptiveBFrame Boolean
    Whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    entropyMode String
    The entropy mode to be used for this layer. If not specified, the encoder chooses the mode that is appropriate for the profile and level.
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.264 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String
    We currently support Baseline, Main, High, High422, High444. Default is Auto.
    referenceFrames Number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    H264Video, H264VideoArgs

    Complexity string | Pulumi.AzureNative.Media.H264Complexity
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.H264Layer>
    The collection of output H.264 layers to be produced by the encoder.
    SceneChangeDetection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    Complexity string | H264Complexity
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []H264Layer
    The collection of output H.264 layers to be produced by the encoder.
    SceneChangeDetection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    complexity String | H264Complexity
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<H264Layer>
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    complexity string | H264Complexity
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers H264Layer[]
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    complexity str | H264Complexity
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[H264Layer]
    The collection of output H.264 layers to be produced by the encoder.
    scene_change_detection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    complexity String | "Speed" | "Balanced" | "Quality"
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    H264VideoProfile, H264VideoProfileArgs

    Auto
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    Baseline
    BaselineBaseline profile
    Main
    MainMain profile
    High
    HighHigh profile.
    High422
    High422High 4:2:2 profile.
    High444
    High444High 4:4:4 predictive profile.
    H264VideoProfileAuto
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    H264VideoProfileBaseline
    BaselineBaseline profile
    H264VideoProfileMain
    MainMain profile
    H264VideoProfileHigh
    HighHigh profile.
    H264VideoProfileHigh422
    High422High 4:2:2 profile.
    H264VideoProfileHigh444
    High444High 4:4:4 predictive profile.
    Auto
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    Baseline
    BaselineBaseline profile
    Main
    MainMain profile
    High
    HighHigh profile.
    High422
    High422High 4:2:2 profile.
    High444
    High444High 4:4:4 predictive profile.
    Auto
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    Baseline
    BaselineBaseline profile
    Main
    MainMain profile
    High
    HighHigh profile.
    High422
    High422High 4:2:2 profile.
    High444
    High444High 4:4:4 predictive profile.
    AUTO
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    BASELINE
    BaselineBaseline profile
    MAIN
    MainMain profile
    HIGH
    HighHigh profile.
    HIGH422
    High422High 4:2:2 profile.
    HIGH444
    High444High 4:4:4 predictive profile.
    "Auto"
    AutoTells the encoder to automatically determine the appropriate H.264 profile.
    "Baseline"
    BaselineBaseline profile
    "Main"
    MainMain profile
    "High"
    HighHigh profile.
    "High422"
    High422High 4:2:2 profile.
    "High444"
    High444High 4:4:4 predictive profile.

    H264VideoResponse, H264VideoResponseArgs

    Complexity string
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.H264LayerResponse>
    The collection of output H.264 layers to be produced by the encoder.
    SceneChangeDetection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    Complexity string
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []H264LayerResponse
    The collection of output H.264 layers to be produced by the encoder.
    SceneChangeDetection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    complexity String
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<H264LayerResponse>
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    complexity string
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers H264LayerResponse[]
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    complexity str
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[H264LayerResponse]
    The collection of output H.264 layers to be produced by the encoder.
    scene_change_detection bool
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    complexity String
    Tells the encoder how to choose its encoding settings. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    The collection of output H.264 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    H265Complexity, H265ComplexityArgs

    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    H265ComplexitySpeed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    H265ComplexityBalanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    H265ComplexityQuality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    Speed
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    Balanced
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    Quality
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    SPEED
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    BALANCED
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    QUALITY
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.
    "Speed"
    SpeedTells the encoder to use settings that are optimized for faster encoding. Quality is sacrificed to decrease encoding time.
    "Balanced"
    BalancedTells the encoder to use settings that achieve a balance between speed and quality.
    "Quality"
    QualityTells the encoder to use settings that are optimized to produce higher quality output at the expense of slower overall encode time.

    H265Layer, H265LayerArgs

    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    AdaptiveBFrame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string | Pulumi.AzureNative.Media.H265VideoProfile
    We currently support Main. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    AdaptiveBFrame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string | H265VideoProfile
    We currently support Main. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Integer
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame Boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Integer
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Integer
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String | H265VideoProfile
    We currently support Main. Default is Auto.
    referenceFrames Integer
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Integer
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate number
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile string | H265VideoProfile
    We currently support Main. Default is Auto.
    referenceFrames number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptive_b_frame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    b_frames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    buffer_window str
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frame_rate str
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level str
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    max_bitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile str | H265VideoProfile
    We currently support Main. Default is Auto.
    reference_frames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Number
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame Boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String | "Auto" | "Main"
    We currently support Main. Default is Auto.
    referenceFrames Number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    H265LayerResponse, H265LayerResponseArgs

    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    AdaptiveBFrame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string
    We currently support Main. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    AdaptiveBFrame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    BFrames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    BufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    FrameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    MaxBitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    Profile string
    We currently support Main. Default is Auto.
    ReferenceFrames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    Slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Integer
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame Boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Integer
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Integer
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String
    We currently support Main. Default is Auto.
    referenceFrames Integer
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Integer
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate number
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow string
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate string
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level string
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile string
    We currently support Main. Default is Auto.
    referenceFrames number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate int
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptive_b_frame bool
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    b_frames int
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    buffer_window str
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frame_rate str
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level str
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    max_bitrate int
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile str
    We currently support Main. Default is Auto.
    reference_frames int
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices int
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    bitrate Number
    The average bitrate in bits per second at which to encode the input video when generating this layer. For example: a target bitrate of 3000Kbps or 3Mbps means this value should be 3000000 This is a required field.
    adaptiveBFrame Boolean
    Specifies whether or not adaptive B-frames are to be used when encoding this layer. If not specified, the encoder will turn it on whenever the video profile permits its use.
    bFrames Number
    The number of B-frames to be used when encoding this layer. If not specified, the encoder chooses an appropriate number based on the video profile and level.
    bufferWindow String
    The VBV buffer window length. The value should be in ISO 8601 format. The value should be in the range [0.1-100] seconds. The default is 5 seconds (for example, PT5S).
    frameRate String
    The frame rate (in frames per second) at which to encode this layer. The value can be in the form of M/N where M and N are integers (For example, 30000/1001), or in the form of a number (For example, 30, or 29.97). The encoder enforces constraints on allowed frame rates based on the profile and level. If it is not specified, the encoder will use the same frame rate as the input video.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    level String
    We currently support Level up to 6.2. The value can be Auto, or a number that matches the H.265 profile. If not specified, the default is Auto, which lets the encoder choose the Level that is appropriate for this layer.
    maxBitrate Number
    The maximum bitrate (in bits per second), at which the VBV buffer should be assumed to refill. If not specified, defaults to the same value as bitrate.
    profile String
    We currently support Main. Default is Auto.
    referenceFrames Number
    The number of reference frames to be used when encoding this layer. If not specified, the encoder determines an appropriate number based on the encoder complexity setting.
    slices Number
    The number of slices to be used when encoding this layer. If not specified, default is zero, which means that encoder will use a single slice for each frame.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    H265Video, H265VideoArgs

    Complexity string | Pulumi.AzureNative.Media.H265Complexity
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.H265Layer>
    The collection of output H.265 layers to be produced by the encoder.
    SceneChangeDetection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    Complexity string | H265Complexity
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []H265Layer
    The collection of output H.265 layers to be produced by the encoder.
    SceneChangeDetection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    complexity String | H265Complexity
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<H265Layer>
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    complexity string | H265Complexity
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers H265Layer[]
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    complexity str | H265Complexity
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[H265Layer]
    The collection of output H.265 layers to be produced by the encoder.
    scene_change_detection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    complexity String | "Speed" | "Balanced" | "Quality"
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    H265VideoProfile, H265VideoProfileArgs

    Auto
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    Main
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
    H265VideoProfileAuto
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    H265VideoProfileMain
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
    Auto
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    Main
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
    Auto
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    Main
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
    AUTO
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    MAIN
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)
    "Auto"
    AutoTells the encoder to automatically determine the appropriate H.265 profile.
    "Main"
    MainMain profile (https://x265.readthedocs.io/en/default/cli.html?highlight=profile#profile-level-tier)

    H265VideoResponse, H265VideoResponseArgs

    Complexity string
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.H265LayerResponse>
    The collection of output H.265 layers to be produced by the encoder.
    SceneChangeDetection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    Complexity string
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []H265LayerResponse
    The collection of output H.265 layers to be produced by the encoder.
    SceneChangeDetection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    complexity String
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<H265LayerResponse>
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    complexity string
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers H265LayerResponse[]
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    complexity str
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[H265LayerResponse]
    The collection of output H.265 layers to be produced by the encoder.
    scene_change_detection bool
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    complexity String
    Tells the encoder how to choose its encoding settings. Quality will provide for a higher compression ratio but at a higher cost and longer compute time. Speed will produce a relatively larger file but is faster and more economical. The default value is Balanced.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    The collection of output H.265 layers to be produced by the encoder.
    sceneChangeDetection Boolean
    Specifies whether or not the encoder should insert key frames at scene changes. If not specified, the default is false. This flag should be set to true only when the encoder is being configured to produce a single output video.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    Image, ImageArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    ImageFormat, ImageFormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    ImageFormatResponse, ImageFormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    ImageResponse, ImageResponseArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    InsightsType, InsightsTypeArgs

    AudioInsightsOnly
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    VideoInsightsOnly
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    AllInsights
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
    InsightsTypeAudioInsightsOnly
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    InsightsTypeVideoInsightsOnly
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    InsightsTypeAllInsights
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
    AudioInsightsOnly
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    VideoInsightsOnly
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    AllInsights
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
    AudioInsightsOnly
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    VideoInsightsOnly
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    AllInsights
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
    AUDIO_INSIGHTS_ONLY
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    VIDEO_INSIGHTS_ONLY
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    ALL_INSIGHTS
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.
    "AudioInsightsOnly"
    AudioInsightsOnlyGenerate audio only insights. Ignore video even if present. Fails if no audio is present.
    "VideoInsightsOnly"
    VideoInsightsOnlyGenerate video only insights. Ignore audio if present. Fails if no video is present.
    "AllInsights"
    AllInsightsGenerate both audio and video insights. Fails if either audio or video Insights fail.

    JpgFormat, JpgFormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    JpgFormatResponse, JpgFormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    JpgImage, JpgImageArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayer>
    A collection of output JPEG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    SpriteColumn int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []JpgLayer
    A collection of output JPEG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    SpriteColumn int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<JpgLayer>
    A collection of output JPEG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn Integer
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers JpgLayer[]
    A collection of output JPEG image layers to be produced by the encoder.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn number
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[JpgLayer]
    A collection of output JPEG image layers to be produced by the encoder.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    sprite_column int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    A collection of output JPEG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn Number
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    JpgImageResponse, JpgImageResponseArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.JpgLayerResponse>
    A collection of output JPEG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    SpriteColumn int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []JpgLayerResponse
    A collection of output JPEG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    SpriteColumn int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<JpgLayerResponse>
    A collection of output JPEG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn Integer
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers JpgLayerResponse[]
    A collection of output JPEG image layers to be produced by the encoder.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn number
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[JpgLayerResponse]
    A collection of output JPEG image layers to be produced by the encoder.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    sprite_column int
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    A collection of output JPEG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    spriteColumn Number
    Sets the number of columns used in thumbnail sprite image. The number of rows are automatically calculated and a VTT file is generated with the coordinate mappings for each thumbnail in the sprite. Note: this value should be a positive integer and a proper value is recommended so that the output image resolution will not go beyond JPEG maximum pixel resolution limit 65535x65535.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    JpgLayer, JpgLayerArgs

    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality Integer
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality number
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality Number
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    JpgLayerResponse, JpgLayerResponseArgs

    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality Integer
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality number
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality int
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    quality Number
    The compression quality of the JPEG output. Range is from 0-100 and the default is 70.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    Mp4Format, Mp4FormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFile
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFile[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFile]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    Mp4FormatResponse, Mp4FormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFileResponse
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFileResponse[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFileResponse]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    MultiBitrateFormat, MultiBitrateFormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFile
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFile[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFile]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    MultiBitrateFormatResponse, MultiBitrateFormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFileResponse
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFileResponse[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFileResponse]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    OnErrorType, OnErrorTypeArgs

    StopProcessingJob
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    ContinueJob
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
    OnErrorTypeStopProcessingJob
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    OnErrorTypeContinueJob
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
    StopProcessingJob
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    ContinueJob
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
    StopProcessingJob
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    ContinueJob
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
    STOP_PROCESSING_JOB
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    CONTINUE_JOB
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.
    "StopProcessingJob"
    StopProcessingJobTells the service that if this TransformOutput fails, then any other incomplete TransformOutputs can be stopped.
    "ContinueJob"
    ContinueJobTells the service that if this TransformOutput fails, then allow any other TransformOutput to continue.

    OutputFile, OutputFileArgs

    Labels List<string>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    Labels []string
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels List<String>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels string[]
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels Sequence[str]
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels List<String>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

    OutputFileResponse, OutputFileResponseArgs

    Labels List<string>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    Labels []string
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels List<String>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels string[]
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels Sequence[str]
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.
    labels List<String>
    The list of labels that describe how the encoder should multiplex video and audio into an output file. For example, if the encoder is producing two video layers with labels v1 and v2, and one audio layer with label a1, then an array like '[v1, a1]' tells the encoder to produce an output file with the video track represented by v1 and the audio track represented by a1.

    PngFormat, PngFormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    PngFormatResponse, PngFormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.

    PngImage, PngImageArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.PngLayer>
    A collection of output PNG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []PngLayer
    A collection of output PNG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<PngLayer>
    A collection of output PNG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers PngLayer[]
    A collection of output PNG image layers to be produced by the encoder.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[PngLayer]
    A collection of output PNG image layers to be produced by the encoder.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    A collection of output PNG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    PngImageResponse, PngImageResponseArgs

    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers List<Pulumi.AzureNative.Media.Inputs.PngLayerResponse>
    A collection of output PNG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    Start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    Layers []PngLayerResponse
    A collection of output PNG image layers to be produced by the encoder.
    Range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    Step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<PngLayerResponse>
    A collection of output PNG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    start string
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    layers PngLayerResponse[]
    A collection of output PNG image layers to be produced by the encoder.
    range string
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step string
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    start str
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    layers Sequence[PngLayerResponse]
    A collection of output PNG image layers to be produced by the encoder.
    range str
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step str
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    start String
    The position in the input video from where to start generating thumbnails. The value can be in ISO 8601 format (For example, PT05S to start at 5 seconds), or a frame count (For example, 10 to start at the 10th frame), or a relative value to stream duration (For example, 10% to start at 10% of stream duration). Also supports a macro {Best}, which tells the encoder to select the best thumbnail from the first few seconds of the video and will only produce one thumbnail, no matter what other settings are for Step and Range. The default value is macro {Best}.
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    layers List<Property Map>
    A collection of output PNG image layers to be produced by the encoder.
    range String
    The position relative to transform preset start time in the input video at which to stop generating thumbnails. The value can be in ISO 8601 format (For example, PT5M30S to stop at 5 minutes and 30 seconds from start time), or a frame count (For example, 300 to stop at the 300th frame from the frame at start time. If this value is 1, it means only producing one thumbnail at start time), or a relative value to the stream duration (For example, 50% to stop at half of stream duration from start time). The default value is 100%, which means to stop at the end of the stream.
    step String
    The intervals at which thumbnails are generated. The value can be in ISO 8601 format (For example, PT05S for one image every 5 seconds), or a frame count (For example, 30 for one image every 30 frames), or a relative value to stream duration (For example, 10% for one image every 10% of stream duration). Note: Step value will affect the first generated thumbnail, which may not be exactly the one specified at transform preset start time. This is due to the encoder, which tries to select the best thumbnail between start time and Step position from start time as the first output. As the default value is 10%, it means if stream has long duration, the first generated thumbnail might be far away from the one specified at start time. Try to select reasonable value for Step if the first thumbnail is expected close to start time, or set Range value at 1 if only one thumbnail is needed at start time.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    PngLayer, PngLayerArgs

    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    PngLayerResponse, PngLayerResponseArgs

    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    Height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    Label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    Width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height string
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label string
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width string
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height str
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label str
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width str
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.
    height String
    The height of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in height as the input.
    label String
    The alphanumeric label for this layer, which can be used in multiplexing different video and audio layers, or in naming the output file.
    width String
    The width of the output video for this layer. The value can be absolute (in pixels) or relative (in percentage). For example 50% means the output video has half as many pixels in width as the input.

    Priority, PriorityArgs

    Low
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    Normal
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    High
    HighUsed for TransformOutputs that should take precedence over others.
    PriorityLow
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    PriorityNormal
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    PriorityHigh
    HighUsed for TransformOutputs that should take precedence over others.
    Low
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    Normal
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    High
    HighUsed for TransformOutputs that should take precedence over others.
    Low
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    Normal
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    High
    HighUsed for TransformOutputs that should take precedence over others.
    LOW
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    NORMAL
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    HIGH
    HighUsed for TransformOutputs that should take precedence over others.
    "Low"
    LowUsed for TransformOutputs that can be generated after Normal and High priority TransformOutputs.
    "Normal"
    NormalUsed for TransformOutputs that can be generated at Normal priority.
    "High"
    HighUsed for TransformOutputs that should take precedence over others.

    Rectangle, RectangleArgs

    Height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height String
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left String
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top String
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width String
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height str
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left str
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top str
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width str
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height String
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left String
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top String
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width String
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

    RectangleResponse, RectangleResponseArgs

    Height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    Width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height String
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left String
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top String
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width String
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height string
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left string
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top string
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width string
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height str
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left str
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top str
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width str
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    height String
    The height of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    left String
    The number of pixels from the left-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    top String
    The number of pixels from the top-margin. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).
    width String
    The width of the rectangular region in pixels. This can be absolute pixel value (e.g 100), or relative to the size of the video (For example, 50%).

    Rotation, RotationArgs

    Auto
    AutoAutomatically detect and rotate as needed.
    None
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    Rotate0
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    Rotate90
    Rotate90Rotate 90 degrees clockwise.
    Rotate180
    Rotate180Rotate 180 degrees clockwise.
    Rotate270
    Rotate270Rotate 270 degrees clockwise.
    RotationAuto
    AutoAutomatically detect and rotate as needed.
    RotationNone
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    RotationRotate0
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    RotationRotate90
    Rotate90Rotate 90 degrees clockwise.
    RotationRotate180
    Rotate180Rotate 180 degrees clockwise.
    RotationRotate270
    Rotate270Rotate 270 degrees clockwise.
    Auto
    AutoAutomatically detect and rotate as needed.
    None
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    Rotate0
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    Rotate90
    Rotate90Rotate 90 degrees clockwise.
    Rotate180
    Rotate180Rotate 180 degrees clockwise.
    Rotate270
    Rotate270Rotate 270 degrees clockwise.
    Auto
    AutoAutomatically detect and rotate as needed.
    None
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    Rotate0
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    Rotate90
    Rotate90Rotate 90 degrees clockwise.
    Rotate180
    Rotate180Rotate 180 degrees clockwise.
    Rotate270
    Rotate270Rotate 270 degrees clockwise.
    AUTO
    AutoAutomatically detect and rotate as needed.
    NONE
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    ROTATE0
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    ROTATE90
    Rotate90Rotate 90 degrees clockwise.
    ROTATE180
    Rotate180Rotate 180 degrees clockwise.
    ROTATE270
    Rotate270Rotate 270 degrees clockwise.
    "Auto"
    AutoAutomatically detect and rotate as needed.
    "None"
    NoneDo not rotate the video. If the output format supports it, any metadata about rotation is kept intact.
    "Rotate0"
    Rotate0Do not rotate the video but remove any metadata about the rotation.
    "Rotate90"
    Rotate90Rotate 90 degrees clockwise.
    "Rotate180"
    Rotate180Rotate 180 degrees clockwise.
    "Rotate270"
    Rotate270Rotate 270 degrees clockwise.

    StandardEncoderPreset, StandardEncoderPresetArgs

    Codecs List<object>
    The list of codecs to be used when encoding the input video.
    Formats List<object>
    The list of outputs to be produced by the encoder.
    Filters Pulumi.AzureNative.Media.Inputs.Filters
    One or more filtering operations that are applied to the input media before encoding.
    Codecs []interface{}
    The list of codecs to be used when encoding the input video.
    Formats []interface{}
    The list of outputs to be produced by the encoder.
    Filters Filters
    One or more filtering operations that are applied to the input media before encoding.
    codecs List<Object>
    The list of codecs to be used when encoding the input video.
    formats List<Object>
    The list of outputs to be produced by the encoder.
    filters Filters
    One or more filtering operations that are applied to the input media before encoding.
    codecs (AacAudio | Audio | CopyAudio | CopyVideo | H264Video | H265Video | Image | JpgImage | PngImage | Video)[]
    The list of codecs to be used when encoding the input video.
    formats (ImageFormat | JpgFormat | Mp4Format | MultiBitrateFormat | PngFormat | TransportStreamFormat)[]
    The list of outputs to be produced by the encoder.
    filters Filters
    One or more filtering operations that are applied to the input media before encoding.
    codecs Sequence[Union[AacAudio, Audio, CopyAudio, CopyVideo, H264Video, H265Video, Image, JpgImage, PngImage, Video]]
    The list of codecs to be used when encoding the input video.
    formats Sequence[Union[ImageFormat, JpgFormat, Mp4Format, MultiBitrateFormat, PngFormat, TransportStreamFormat]]
    The list of outputs to be produced by the encoder.
    filters Filters
    One or more filtering operations that are applied to the input media before encoding.
    codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
    The list of codecs to be used when encoding the input video.
    formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
    The list of outputs to be produced by the encoder.
    filters Property Map
    One or more filtering operations that are applied to the input media before encoding.

    StandardEncoderPresetResponse, StandardEncoderPresetResponseArgs

    Codecs List<object>
    The list of codecs to be used when encoding the input video.
    Formats List<object>
    The list of outputs to be produced by the encoder.
    Filters Pulumi.AzureNative.Media.Inputs.FiltersResponse
    One or more filtering operations that are applied to the input media before encoding.
    Codecs []interface{}
    The list of codecs to be used when encoding the input video.
    Formats []interface{}
    The list of outputs to be produced by the encoder.
    Filters FiltersResponse
    One or more filtering operations that are applied to the input media before encoding.
    codecs List<Object>
    The list of codecs to be used when encoding the input video.
    formats List<Object>
    The list of outputs to be produced by the encoder.
    filters FiltersResponse
    One or more filtering operations that are applied to the input media before encoding.
    codecs (AacAudioResponse | AudioResponse | CopyAudioResponse | CopyVideoResponse | H264VideoResponse | H265VideoResponse | ImageResponse | JpgImageResponse | PngImageResponse | VideoResponse)[]
    The list of codecs to be used when encoding the input video.
    formats (ImageFormatResponse | JpgFormatResponse | Mp4FormatResponse | MultiBitrateFormatResponse | PngFormatResponse | TransportStreamFormatResponse)[]
    The list of outputs to be produced by the encoder.
    filters FiltersResponse
    One or more filtering operations that are applied to the input media before encoding.
    codecs Sequence[Union[AacAudioResponse, AudioResponse, CopyAudioResponse, CopyVideoResponse, H264VideoResponse, H265VideoResponse, ImageResponse, JpgImageResponse, PngImageResponse, VideoResponse]]
    The list of codecs to be used when encoding the input video.
    formats Sequence[Union[ImageFormatResponse, JpgFormatResponse, Mp4FormatResponse, MultiBitrateFormatResponse, PngFormatResponse, TransportStreamFormatResponse]]
    The list of outputs to be produced by the encoder.
    filters FiltersResponse
    One or more filtering operations that are applied to the input media before encoding.
    codecs List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
    The list of codecs to be used when encoding the input video.
    formats List<Property Map | Property Map | Property Map | Property Map | Property Map | Property Map>
    The list of outputs to be produced by the encoder.
    filters Property Map
    One or more filtering operations that are applied to the input media before encoding.

    StretchMode, StretchModeArgs

    None
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    AutoSize
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    AutoFit
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
    StretchModeNone
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    StretchModeAutoSize
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    StretchModeAutoFit
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
    None
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    AutoSize
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    AutoFit
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
    None
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    AutoSize
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    AutoFit
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
    NONE
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    AUTO_SIZE
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    AUTO_FIT
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.
    "None"
    NoneStrictly respect the output resolution without considering the pixel aspect ratio or display aspect ratio of the input video.
    "AutoSize"
    AutoSizeOverride the output resolution, and change it to match the display aspect ratio of the input, without padding. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the value in the preset is overridden, and the output will be at 1280x720, which maintains the input aspect ratio of 16:9.
    "AutoFit"
    AutoFitPad the output (with either letterbox or pillar box) to honor the output resolution, while ensuring that the active video region in the output has the same aspect ratio as the input. For example, if the input is 1920x1080 and the encoding preset asks for 1280x1280, then the output will be at 1280x1280, which contains an inner rectangle of 1280x720 at aspect ratio of 16:9, and pillar box regions 280 pixels wide at the left and right.

    SystemDataResponse, SystemDataResponseArgs

    CreatedAt string
    The timestamp of resource creation (UTC).
    CreatedBy string
    The identity that created the resource.
    CreatedByType string
    The type of identity that created the resource.
    LastModifiedAt string
    The timestamp of resource last modification (UTC)
    LastModifiedBy string
    The identity that last modified the resource.
    LastModifiedByType string
    The type of identity that last modified the resource.
    CreatedAt string
    The timestamp of resource creation (UTC).
    CreatedBy string
    The identity that created the resource.
    CreatedByType string
    The type of identity that created the resource.
    LastModifiedAt string
    The timestamp of resource last modification (UTC)
    LastModifiedBy string
    The identity that last modified the resource.
    LastModifiedByType string
    The type of identity that last modified the resource.
    createdAt String
    The timestamp of resource creation (UTC).
    createdBy String
    The identity that created the resource.
    createdByType String
    The type of identity that created the resource.
    lastModifiedAt String
    The timestamp of resource last modification (UTC)
    lastModifiedBy String
    The identity that last modified the resource.
    lastModifiedByType String
    The type of identity that last modified the resource.
    createdAt string
    The timestamp of resource creation (UTC).
    createdBy string
    The identity that created the resource.
    createdByType string
    The type of identity that created the resource.
    lastModifiedAt string
    The timestamp of resource last modification (UTC)
    lastModifiedBy string
    The identity that last modified the resource.
    lastModifiedByType string
    The type of identity that last modified the resource.
    created_at str
    The timestamp of resource creation (UTC).
    created_by str
    The identity that created the resource.
    created_by_type str
    The type of identity that created the resource.
    last_modified_at str
    The timestamp of resource last modification (UTC)
    last_modified_by str
    The identity that last modified the resource.
    last_modified_by_type str
    The type of identity that last modified the resource.
    createdAt String
    The timestamp of resource creation (UTC).
    createdBy String
    The identity that created the resource.
    createdByType String
    The type of identity that created the resource.
    lastModifiedAt String
    The timestamp of resource last modification (UTC)
    lastModifiedBy String
    The identity that last modified the resource.
    lastModifiedByType String
    The type of identity that last modified the resource.

    TransformOutput, TransformOutputArgs

    Preset Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPreset | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.FaceDetectorPreset | Pulumi.AzureNative.Media.Inputs.StandardEncoderPreset | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPreset
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    OnError string | Pulumi.AzureNative.Media.OnErrorType
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    RelativePriority string | Pulumi.AzureNative.Media.Priority
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    Preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    OnError string | OnErrorType
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    RelativePriority string | Priority
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError String | OnErrorType
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority String | Priority
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError string | OnErrorType
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority string | Priority
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPreset | BuiltInStandardEncoderPreset | FaceDetectorPreset | StandardEncoderPreset | VideoAnalyzerPreset
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    on_error str | OnErrorType
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relative_priority str | Priority
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset Property Map | Property Map | Property Map | Property Map | Property Map
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError String | "StopProcessingJob" | "ContinueJob"
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority String | "Low" | "Normal" | "High"
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

    TransformOutputResponse, TransformOutputResponseArgs

    Preset Pulumi.AzureNative.Media.Inputs.AudioAnalyzerPresetResponse | Pulumi.AzureNative.Media.Inputs.BuiltInStandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.FaceDetectorPresetResponse | Pulumi.AzureNative.Media.Inputs.StandardEncoderPresetResponse | Pulumi.AzureNative.Media.Inputs.VideoAnalyzerPresetResponse
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    OnError string
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    RelativePriority string
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    Preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    OnError string
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    RelativePriority string
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError String
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority String
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError string
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority string
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset AudioAnalyzerPresetResponse | BuiltInStandardEncoderPresetResponse | FaceDetectorPresetResponse | StandardEncoderPresetResponse | VideoAnalyzerPresetResponse
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    on_error str
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relative_priority str
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.
    preset Property Map | Property Map | Property Map | Property Map | Property Map
    Preset that describes the operations that will be used to modify, transcode, or extract insights from the source file to generate the output.
    onError String
    A Transform can define more than one outputs. This property defines what the service should do when one output fails - either continue to produce other outputs, or, stop the other outputs. The overall Job state will not reflect failures of outputs that are specified with 'ContinueJob'. The default is 'StopProcessingJob'.
    relativePriority String
    Sets the relative priority of the TransformOutputs within a Transform. This sets the priority that the service uses for processing TransformOutputs. The default priority is Normal.

    TransportStreamFormat, TransportStreamFormatArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFile
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFile>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFile[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFile]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    TransportStreamFormatResponse, TransportStreamFormatResponseArgs

    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles List<Pulumi.AzureNative.Media.Inputs.OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    FilenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    OutputFiles []OutputFileResponse
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<OutputFileResponse>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern string
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles OutputFileResponse[]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filename_pattern str
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    output_files Sequence[OutputFileResponse]
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .
    filenamePattern String
    The pattern of the file names for the generated output files. The following macros are supported in the file name: {Basename} - An expansion macro that will use the name of the input video file. If the base name(the file suffix is not included) of the input video file is less than 32 characters long, the base name of input video files will be used. If the length of base name of the input video file exceeds 32 characters, the base name is truncated to the first 32 characters in total length. {Extension} - The appropriate extension for this format. {Label} - The label assigned to the codec/layer. {Index} - A unique index for thumbnails. Only applicable to thumbnails. {Bitrate} - The audio/video bitrate. Not applicable to thumbnails. {Codec} - The type of the audio/video codec. {Resolution} - The video resolution. Any unsubstituted macros will be collapsed and removed from the filename.
    outputFiles List<Property Map>
    The list of output files to produce. Each entry in the list is a set of audio and video layer labels to be muxed together .

    Video, VideoArgs

    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    StretchMode string | Pulumi.AzureNative.Media.StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | Pulumi.AzureNative.Media.VideoSyncMode
    The Video Sync Mode
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    StretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string | VideoSyncMode
    The Video Sync Mode
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode String | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | VideoSyncMode
    The Video Sync Mode
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode string | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string | VideoSyncMode
    The Video Sync Mode
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    stretch_mode str | StretchMode
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str | VideoSyncMode
    The Video Sync Mode
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode String | "None" | "AutoSize" | "AutoFit"
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String | "Auto" | "Passthrough" | "Cfr" | "Vfr"
    The Video Sync Mode

    VideoAnalyzerPreset, VideoAnalyzerPresetArgs

    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    InsightsToExtract string | Pulumi.AzureNative.Media.InsightsType
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    Mode string | Pulumi.AzureNative.Media.AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    InsightsToExtract string | InsightsType
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    Mode string | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract String | InsightsType
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode String | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract string | InsightsType
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode string | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audio_language str
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insights_to_extract str | InsightsType
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode str | AudioAnalysisMode
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract String | "AudioInsightsOnly" | "VideoInsightsOnly" | "AllInsights"
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode String | "Standard" | "Basic"
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

    VideoAnalyzerPresetResponse, VideoAnalyzerPresetResponseArgs

    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions Dictionary<string, string>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    InsightsToExtract string
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    Mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    AudioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    ExperimentalOptions map[string]string
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    InsightsToExtract string
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    Mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String,String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract String
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode String
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage string
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions {[key: string]: string}
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract string
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode string
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audio_language str
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimental_options Mapping[str, str]
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insights_to_extract str
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode str
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.
    audioLanguage String
    The language for the audio payload in the input using the BCP-47 format of 'language tag-region' (e.g: 'en-US'). If you know the language of your content, it is recommended that you specify it. The language must be specified explicitly for AudioAnalysisMode::Basic, since automatic language detection is not included in basic mode. If the language isn't specified or set to null, automatic language detection will choose the first language detected and process with the selected language for the duration of the file. It does not currently support dynamically switching between languages after the first language is detected. The automatic detection works best with audio recordings with clearly discernable speech. If automatic detection fails to find the language, transcription would fallback to 'en-US'." The list of supported languages is available here: https://go.microsoft.com/fwlink/?linkid=2109463
    experimentalOptions Map<String>
    Dictionary containing key value pairs for parameters not exposed in the preset itself
    insightsToExtract String
    Defines the type of insights that you want the service to generate. The allowed values are 'AudioInsightsOnly', 'VideoInsightsOnly', and 'AllInsights'. The default is AllInsights. If you set this to AllInsights and the input is audio only, then only audio insights are generated. Similarly if the input is video only, then only video insights are generated. It is recommended that you not use AudioInsightsOnly if you expect some of your inputs to be video only; or use VideoInsightsOnly if you expect some of your inputs to be audio only. Your Jobs in such conditions would error out.
    mode String
    Determines the set of audio analysis operations to be performed. If unspecified, the Standard AudioAnalysisMode would be chosen.

    VideoOverlay, VideoOverlayArgs

    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    CropRectangle Pulumi.AzureNative.Media.Inputs.Rectangle
    An optional rectangular window used to crop the overlay image or video.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Opacity double
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    Position Pulumi.AzureNative.Media.Inputs.Rectangle
    The location in the input video where the overlay is applied.
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel float64
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    CropRectangle Rectangle
    An optional rectangular window used to crop the overlay image or video.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Opacity float64
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    Position Rectangle
    The location in the input video where the overlay is applied.
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle Rectangle
    An optional rectangular window used to crop the overlay image or video.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity Double
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position Rectangle
    The location in the input video where the overlay is applied.
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle Rectangle
    An optional rectangular window used to crop the overlay image or video.
    end string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity number
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position Rectangle
    The location in the input video where the overlay is applied.
    start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    input_label str
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audio_gain_level float
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    crop_rectangle Rectangle
    An optional rectangular window used to crop the overlay image or video.
    end str
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fade_in_duration str
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fade_out_duration str
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity float
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position Rectangle
    The location in the input video where the overlay is applied.
    start str
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle Property Map
    An optional rectangular window used to crop the overlay image or video.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity Number
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position Property Map
    The location in the input video where the overlay is applied.
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

    VideoOverlayResponse, VideoOverlayResponseArgs

    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    CropRectangle Pulumi.AzureNative.Media.Inputs.RectangleResponse
    An optional rectangular window used to crop the overlay image or video.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Opacity double
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    Position Pulumi.AzureNative.Media.Inputs.RectangleResponse
    The location in the input video where the overlay is applied.
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    InputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    AudioGainLevel float64
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    CropRectangle RectangleResponse
    An optional rectangular window used to crop the overlay image or video.
    End string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    FadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    FadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    Opacity float64
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    Position RectangleResponse
    The location in the input video where the overlay is applied.
    Start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Double
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle RectangleResponse
    An optional rectangular window used to crop the overlay image or video.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity Double
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position RectangleResponse
    The location in the input video where the overlay is applied.
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel string
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle RectangleResponse
    An optional rectangular window used to crop the overlay image or video.
    end string
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration string
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration string
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity number
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position RectangleResponse
    The location in the input video where the overlay is applied.
    start string
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    input_label str
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audio_gain_level float
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    crop_rectangle RectangleResponse
    An optional rectangular window used to crop the overlay image or video.
    end str
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fade_in_duration str
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fade_out_duration str
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity float
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position RectangleResponse
    The location in the input video where the overlay is applied.
    start str
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.
    inputLabel String
    The label of the job input which is to be used as an overlay. The Input must specify exactly one file. You can specify an image file in JPG, PNG, GIF or BMP format, or an audio file (such as a WAV, MP3, WMA or M4A file), or a video file. See https://aka.ms/mesformats for the complete list of supported audio and video file formats.
    audioGainLevel Number
    The gain level of audio in the overlay. The value should be in the range [0, 1.0]. The default is 1.0.
    cropRectangle Property Map
    An optional rectangular window used to crop the overlay image or video.
    end String
    The end position, with reference to the input video, at which the overlay ends. The value should be in ISO 8601 format. For example, PT30S to end the overlay at 30 seconds into the input video. If not specified or the value is greater than the input video duration, the overlay will be applied until the end of the input video if the overlay media duration is greater than the input video duration, else the overlay will last as long as the overlay media duration.
    fadeInDuration String
    The duration over which the overlay fades in onto the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade in (same as PT0S).
    fadeOutDuration String
    The duration over which the overlay fades out of the input video. The value should be in ISO 8601 duration format. If not specified the default behavior is to have no fade out (same as PT0S).
    opacity Number
    The opacity of the overlay. This is a value in the range [0 - 1.0]. Default is 1.0 which mean the overlay is opaque.
    position Property Map
    The location in the input video where the overlay is applied.
    start String
    The start position, with reference to the input video, at which the overlay starts. The value should be in ISO 8601 format. For example, PT05S to start the overlay at 5 seconds into the input video. If not specified the overlay starts from the beginning of the input video.

    VideoResponse, VideoResponseArgs

    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    KeyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    Label string
    An optional label for the codec. The label can be used to control muxing behavior.
    StretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    SyncMode string
    The Video Sync Mode
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode
    keyFrameInterval string
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label string
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode string
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode string
    The Video Sync Mode
    key_frame_interval str
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label str
    An optional label for the codec. The label can be used to control muxing behavior.
    stretch_mode str
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    sync_mode str
    The Video Sync Mode
    keyFrameInterval String
    The distance between two key frames. The value should be non-zero in the range [0.5, 20] seconds, specified in ISO 8601 format. The default is 2 seconds(PT2S). Note that this setting is ignored if VideoSyncMode.Passthrough is set, where the KeyFrameInterval value will follow the input source setting.
    label String
    An optional label for the codec. The label can be used to control muxing behavior.
    stretchMode String
    The resizing mode - how the input video will be resized to fit the desired output resolution(s). Default is AutoSize
    syncMode String
    The Video Sync Mode

    VideoSyncMode, VideoSyncModeArgs

    Auto
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    Passthrough
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    Cfr
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    Vfr
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
    VideoSyncModeAuto
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    VideoSyncModePassthrough
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    VideoSyncModeCfr
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    VideoSyncModeVfr
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
    Auto
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    Passthrough
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    Cfr
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    Vfr
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
    Auto
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    Passthrough
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    Cfr
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    Vfr
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
    AUTO
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    PASSTHROUGH
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    CFR
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    VFR
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input
    "Auto"
    AutoThis is the default method. Chooses between Cfr and Vfr depending on muxer capabilities. For output format MP4, the default mode is Cfr.
    "Passthrough"
    PassthroughThe presentation timestamps on frames are passed through from the input file to the output file writer. Recommended when the input source has variable frame rate, and are attempting to produce multiple layers for adaptive streaming in the output which have aligned GOP boundaries. Note: if two or more frames in the input have duplicate timestamps, then the output will also have the same behavior
    "Cfr"
    CfrInput frames will be repeated and/or dropped as needed to achieve exactly the requested constant frame rate. Recommended when the output frame rate is explicitly set at a specified value
    "Vfr"
    VfrSimilar to the Passthrough mode, but if the input has frames that have duplicate timestamps, then only one frame is passed through to the output, and others are dropped. Recommended when the number of output frames is expected to be equal to the number of input frames. For example, the output is used to calculate a quality metric like PSNR against the input

    Import

    An existing resource can be imported using its type token, name, and identifier, e.g.

    $ pulumi import azure-native:media:Transform createdTransform /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/contosoresources/providers/Microsoft.Media/mediaservices/contosomedia/transforms/createdTransform 
    

    To learn more about importing existing cloud resources, see Importing resources.

    Package Details

    Repository
    azure-native-v1 pulumi/pulumi-azure-native
    License
    Apache-2.0
    azure-native-v1 logo
    These are the docs for Azure Native v1. We recommenend using the latest version, Azure Native v2.
    Azure Native v1 v1.104.0 published on Thursday, Jul 6, 2023 by Pulumi