1. Packages
  2. Azure Classic
  3. API Docs
  4. streamanalytics
  5. JobSchedule

We recommend using Azure Native.

Azure Classic v5.81.0 published on Monday, Jun 24, 2024 by Pulumi

azure.streamanalytics.JobSchedule

Explore with Pulumi AI

azure logo

We recommend using Azure Native.

Azure Classic v5.81.0 published on Monday, Jun 24, 2024 by Pulumi

    Manages a Stream Analytics Job Schedule.

    Example Usage

    import * as pulumi from "@pulumi/pulumi";
    import * as azure from "@pulumi/azure";
    
    const example = new azure.core.ResourceGroup("example", {
        name: "example-resources",
        location: "West Europe",
    });
    const exampleAccount = new azure.storage.Account("example", {
        name: "example",
        resourceGroupName: example.name,
        location: example.location,
        accountTier: "Standard",
        accountReplicationType: "LRS",
    });
    const exampleContainer = new azure.storage.Container("example", {
        name: "example",
        storageAccountName: exampleAccount.name,
        containerAccessType: "private",
    });
    const exampleBlob = new azure.storage.Blob("example", {
        name: "example",
        storageAccountName: exampleAccount.name,
        storageContainerName: exampleContainer.name,
        type: "Block",
        source: new pulumi.asset.FileAsset("example.csv"),
    });
    const exampleJob = new azure.streamanalytics.Job("example", {
        name: "example-job",
        resourceGroupName: example.name,
        location: example.location,
        compatibilityLevel: "1.2",
        dataLocale: "en-GB",
        eventsLateArrivalMaxDelayInSeconds: 60,
        eventsOutOfOrderMaxDelayInSeconds: 50,
        eventsOutOfOrderPolicy: "Adjust",
        outputErrorPolicy: "Drop",
        streamingUnits: 3,
        tags: {
            environment: "Example",
        },
        transformationQuery: `    SELECT *
        INTO [exampleoutput]
        FROM [exampleinput]
    `,
    });
    const exampleStreamInputBlob = new azure.streamanalytics.StreamInputBlob("example", {
        name: "exampleinput",
        streamAnalyticsJobName: exampleJob.name,
        resourceGroupName: exampleJob.resourceGroupName,
        storageAccountName: exampleAccount.name,
        storageAccountKey: exampleAccount.primaryAccessKey,
        storageContainerName: exampleContainer.name,
        pathPattern: "",
        dateFormat: "yyyy/MM/dd",
        timeFormat: "HH",
        serialization: {
            type: "Csv",
            encoding: "UTF8",
            fieldDelimiter: ",",
        },
    });
    const exampleOutputBlob = new azure.streamanalytics.OutputBlob("example", {
        name: "exampleoutput",
        streamAnalyticsJobName: exampleJob.name,
        resourceGroupName: exampleJob.resourceGroupName,
        storageAccountName: exampleAccount.name,
        storageAccountKey: exampleAccount.primaryAccessKey,
        storageContainerName: exampleContainer.name,
        pathPattern: "example-{date}-{time}",
        dateFormat: "yyyy-MM-dd",
        timeFormat: "HH",
        serialization: {
            type: "Avro",
        },
    });
    const exampleJobSchedule = new azure.streamanalytics.JobSchedule("example", {
        streamAnalyticsJobId: exampleJob.id,
        startMode: "CustomTime",
        startTime: "2022-09-21T00:00:00Z",
    }, {
        dependsOn: [
            exampleJob,
            exampleStreamInputBlob,
            exampleOutputBlob,
        ],
    });
    
    import pulumi
    import pulumi_azure as azure
    
    example = azure.core.ResourceGroup("example",
        name="example-resources",
        location="West Europe")
    example_account = azure.storage.Account("example",
        name="example",
        resource_group_name=example.name,
        location=example.location,
        account_tier="Standard",
        account_replication_type="LRS")
    example_container = azure.storage.Container("example",
        name="example",
        storage_account_name=example_account.name,
        container_access_type="private")
    example_blob = azure.storage.Blob("example",
        name="example",
        storage_account_name=example_account.name,
        storage_container_name=example_container.name,
        type="Block",
        source=pulumi.FileAsset("example.csv"))
    example_job = azure.streamanalytics.Job("example",
        name="example-job",
        resource_group_name=example.name,
        location=example.location,
        compatibility_level="1.2",
        data_locale="en-GB",
        events_late_arrival_max_delay_in_seconds=60,
        events_out_of_order_max_delay_in_seconds=50,
        events_out_of_order_policy="Adjust",
        output_error_policy="Drop",
        streaming_units=3,
        tags={
            "environment": "Example",
        },
        transformation_query="""    SELECT *
        INTO [exampleoutput]
        FROM [exampleinput]
    """)
    example_stream_input_blob = azure.streamanalytics.StreamInputBlob("example",
        name="exampleinput",
        stream_analytics_job_name=example_job.name,
        resource_group_name=example_job.resource_group_name,
        storage_account_name=example_account.name,
        storage_account_key=example_account.primary_access_key,
        storage_container_name=example_container.name,
        path_pattern="",
        date_format="yyyy/MM/dd",
        time_format="HH",
        serialization=azure.streamanalytics.StreamInputBlobSerializationArgs(
            type="Csv",
            encoding="UTF8",
            field_delimiter=",",
        ))
    example_output_blob = azure.streamanalytics.OutputBlob("example",
        name="exampleoutput",
        stream_analytics_job_name=example_job.name,
        resource_group_name=example_job.resource_group_name,
        storage_account_name=example_account.name,
        storage_account_key=example_account.primary_access_key,
        storage_container_name=example_container.name,
        path_pattern="example-{date}-{time}",
        date_format="yyyy-MM-dd",
        time_format="HH",
        serialization=azure.streamanalytics.OutputBlobSerializationArgs(
            type="Avro",
        ))
    example_job_schedule = azure.streamanalytics.JobSchedule("example",
        stream_analytics_job_id=example_job.id,
        start_mode="CustomTime",
        start_time="2022-09-21T00:00:00Z",
        opts=pulumi.ResourceOptions(depends_on=[
                example_job,
                example_stream_input_blob,
                example_output_blob,
            ]))
    
    package main
    
    import (
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/core"
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/storage"
    	"github.com/pulumi/pulumi-azure/sdk/v5/go/azure/streamanalytics"
    	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
    )
    
    func main() {
    	pulumi.Run(func(ctx *pulumi.Context) error {
    		example, err := core.NewResourceGroup(ctx, "example", &core.ResourceGroupArgs{
    			Name:     pulumi.String("example-resources"),
    			Location: pulumi.String("West Europe"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleAccount, err := storage.NewAccount(ctx, "example", &storage.AccountArgs{
    			Name:                   pulumi.String("example"),
    			ResourceGroupName:      example.Name,
    			Location:               example.Location,
    			AccountTier:            pulumi.String("Standard"),
    			AccountReplicationType: pulumi.String("LRS"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleContainer, err := storage.NewContainer(ctx, "example", &storage.ContainerArgs{
    			Name:                pulumi.String("example"),
    			StorageAccountName:  exampleAccount.Name,
    			ContainerAccessType: pulumi.String("private"),
    		})
    		if err != nil {
    			return err
    		}
    		_, err = storage.NewBlob(ctx, "example", &storage.BlobArgs{
    			Name:                 pulumi.String("example"),
    			StorageAccountName:   exampleAccount.Name,
    			StorageContainerName: exampleContainer.Name,
    			Type:                 pulumi.String("Block"),
    			Source:               pulumi.NewFileAsset("example.csv"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleJob, err := streamanalytics.NewJob(ctx, "example", &streamanalytics.JobArgs{
    			Name:                               pulumi.String("example-job"),
    			ResourceGroupName:                  example.Name,
    			Location:                           example.Location,
    			CompatibilityLevel:                 pulumi.String("1.2"),
    			DataLocale:                         pulumi.String("en-GB"),
    			EventsLateArrivalMaxDelayInSeconds: pulumi.Int(60),
    			EventsOutOfOrderMaxDelayInSeconds:  pulumi.Int(50),
    			EventsOutOfOrderPolicy:             pulumi.String("Adjust"),
    			OutputErrorPolicy:                  pulumi.String("Drop"),
    			StreamingUnits:                     pulumi.Int(3),
    			Tags: pulumi.StringMap{
    				"environment": pulumi.String("Example"),
    			},
    			TransformationQuery: pulumi.String("    SELECT *\n    INTO [exampleoutput]\n    FROM [exampleinput]\n"),
    		})
    		if err != nil {
    			return err
    		}
    		exampleStreamInputBlob, err := streamanalytics.NewStreamInputBlob(ctx, "example", &streamanalytics.StreamInputBlobArgs{
    			Name:                   pulumi.String("exampleinput"),
    			StreamAnalyticsJobName: exampleJob.Name,
    			ResourceGroupName:      exampleJob.ResourceGroupName,
    			StorageAccountName:     exampleAccount.Name,
    			StorageAccountKey:      exampleAccount.PrimaryAccessKey,
    			StorageContainerName:   exampleContainer.Name,
    			PathPattern:            pulumi.String(""),
    			DateFormat:             pulumi.String("yyyy/MM/dd"),
    			TimeFormat:             pulumi.String("HH"),
    			Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
    				Type:           pulumi.String("Csv"),
    				Encoding:       pulumi.String("UTF8"),
    				FieldDelimiter: pulumi.String(","),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		exampleOutputBlob, err := streamanalytics.NewOutputBlob(ctx, "example", &streamanalytics.OutputBlobArgs{
    			Name:                   pulumi.String("exampleoutput"),
    			StreamAnalyticsJobName: exampleJob.Name,
    			ResourceGroupName:      exampleJob.ResourceGroupName,
    			StorageAccountName:     exampleAccount.Name,
    			StorageAccountKey:      exampleAccount.PrimaryAccessKey,
    			StorageContainerName:   exampleContainer.Name,
    			PathPattern:            pulumi.String("example-{date}-{time}"),
    			DateFormat:             pulumi.String("yyyy-MM-dd"),
    			TimeFormat:             pulumi.String("HH"),
    			Serialization: &streamanalytics.OutputBlobSerializationArgs{
    				Type: pulumi.String("Avro"),
    			},
    		})
    		if err != nil {
    			return err
    		}
    		_, err = streamanalytics.NewJobSchedule(ctx, "example", &streamanalytics.JobScheduleArgs{
    			StreamAnalyticsJobId: exampleJob.ID(),
    			StartMode:            pulumi.String("CustomTime"),
    			StartTime:            pulumi.String("2022-09-21T00:00:00Z"),
    		}, pulumi.DependsOn([]pulumi.Resource{
    			exampleJob,
    			exampleStreamInputBlob,
    			exampleOutputBlob,
    		}))
    		if err != nil {
    			return err
    		}
    		return nil
    	})
    }
    
    using System.Collections.Generic;
    using System.Linq;
    using Pulumi;
    using Azure = Pulumi.Azure;
    
    return await Deployment.RunAsync(() => 
    {
        var example = new Azure.Core.ResourceGroup("example", new()
        {
            Name = "example-resources",
            Location = "West Europe",
        });
    
        var exampleAccount = new Azure.Storage.Account("example", new()
        {
            Name = "example",
            ResourceGroupName = example.Name,
            Location = example.Location,
            AccountTier = "Standard",
            AccountReplicationType = "LRS",
        });
    
        var exampleContainer = new Azure.Storage.Container("example", new()
        {
            Name = "example",
            StorageAccountName = exampleAccount.Name,
            ContainerAccessType = "private",
        });
    
        var exampleBlob = new Azure.Storage.Blob("example", new()
        {
            Name = "example",
            StorageAccountName = exampleAccount.Name,
            StorageContainerName = exampleContainer.Name,
            Type = "Block",
            Source = new FileAsset("example.csv"),
        });
    
        var exampleJob = new Azure.StreamAnalytics.Job("example", new()
        {
            Name = "example-job",
            ResourceGroupName = example.Name,
            Location = example.Location,
            CompatibilityLevel = "1.2",
            DataLocale = "en-GB",
            EventsLateArrivalMaxDelayInSeconds = 60,
            EventsOutOfOrderMaxDelayInSeconds = 50,
            EventsOutOfOrderPolicy = "Adjust",
            OutputErrorPolicy = "Drop",
            StreamingUnits = 3,
            Tags = 
            {
                { "environment", "Example" },
            },
            TransformationQuery = @"    SELECT *
        INTO [exampleoutput]
        FROM [exampleinput]
    ",
        });
    
        var exampleStreamInputBlob = new Azure.StreamAnalytics.StreamInputBlob("example", new()
        {
            Name = "exampleinput",
            StreamAnalyticsJobName = exampleJob.Name,
            ResourceGroupName = exampleJob.ResourceGroupName,
            StorageAccountName = exampleAccount.Name,
            StorageAccountKey = exampleAccount.PrimaryAccessKey,
            StorageContainerName = exampleContainer.Name,
            PathPattern = "",
            DateFormat = "yyyy/MM/dd",
            TimeFormat = "HH",
            Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
            {
                Type = "Csv",
                Encoding = "UTF8",
                FieldDelimiter = ",",
            },
        });
    
        var exampleOutputBlob = new Azure.StreamAnalytics.OutputBlob("example", new()
        {
            Name = "exampleoutput",
            StreamAnalyticsJobName = exampleJob.Name,
            ResourceGroupName = exampleJob.ResourceGroupName,
            StorageAccountName = exampleAccount.Name,
            StorageAccountKey = exampleAccount.PrimaryAccessKey,
            StorageContainerName = exampleContainer.Name,
            PathPattern = "example-{date}-{time}",
            DateFormat = "yyyy-MM-dd",
            TimeFormat = "HH",
            Serialization = new Azure.StreamAnalytics.Inputs.OutputBlobSerializationArgs
            {
                Type = "Avro",
            },
        });
    
        var exampleJobSchedule = new Azure.StreamAnalytics.JobSchedule("example", new()
        {
            StreamAnalyticsJobId = exampleJob.Id,
            StartMode = "CustomTime",
            StartTime = "2022-09-21T00:00:00Z",
        }, new CustomResourceOptions
        {
            DependsOn =
            {
                exampleJob,
                exampleStreamInputBlob,
                exampleOutputBlob,
            },
        });
    
    });
    
    package generated_program;
    
    import com.pulumi.Context;
    import com.pulumi.Pulumi;
    import com.pulumi.core.Output;
    import com.pulumi.azure.core.ResourceGroup;
    import com.pulumi.azure.core.ResourceGroupArgs;
    import com.pulumi.azure.storage.Account;
    import com.pulumi.azure.storage.AccountArgs;
    import com.pulumi.azure.storage.Container;
    import com.pulumi.azure.storage.ContainerArgs;
    import com.pulumi.azure.storage.Blob;
    import com.pulumi.azure.storage.BlobArgs;
    import com.pulumi.azure.streamanalytics.Job;
    import com.pulumi.azure.streamanalytics.JobArgs;
    import com.pulumi.azure.streamanalytics.StreamInputBlob;
    import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
    import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
    import com.pulumi.azure.streamanalytics.OutputBlob;
    import com.pulumi.azure.streamanalytics.OutputBlobArgs;
    import com.pulumi.azure.streamanalytics.inputs.OutputBlobSerializationArgs;
    import com.pulumi.azure.streamanalytics.JobSchedule;
    import com.pulumi.azure.streamanalytics.JobScheduleArgs;
    import com.pulumi.resources.CustomResourceOptions;
    import com.pulumi.asset.FileAsset;
    import java.util.List;
    import java.util.ArrayList;
    import java.util.Map;
    import java.io.File;
    import java.nio.file.Files;
    import java.nio.file.Paths;
    
    public class App {
        public static void main(String[] args) {
            Pulumi.run(App::stack);
        }
    
        public static void stack(Context ctx) {
            var example = new ResourceGroup("example", ResourceGroupArgs.builder()
                .name("example-resources")
                .location("West Europe")
                .build());
    
            var exampleAccount = new Account("exampleAccount", AccountArgs.builder()
                .name("example")
                .resourceGroupName(example.name())
                .location(example.location())
                .accountTier("Standard")
                .accountReplicationType("LRS")
                .build());
    
            var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()
                .name("example")
                .storageAccountName(exampleAccount.name())
                .containerAccessType("private")
                .build());
    
            var exampleBlob = new Blob("exampleBlob", BlobArgs.builder()
                .name("example")
                .storageAccountName(exampleAccount.name())
                .storageContainerName(exampleContainer.name())
                .type("Block")
                .source(new FileAsset("example.csv"))
                .build());
    
            var exampleJob = new Job("exampleJob", JobArgs.builder()
                .name("example-job")
                .resourceGroupName(example.name())
                .location(example.location())
                .compatibilityLevel("1.2")
                .dataLocale("en-GB")
                .eventsLateArrivalMaxDelayInSeconds(60)
                .eventsOutOfOrderMaxDelayInSeconds(50)
                .eventsOutOfOrderPolicy("Adjust")
                .outputErrorPolicy("Drop")
                .streamingUnits(3)
                .tags(Map.of("environment", "Example"))
                .transformationQuery("""
        SELECT *
        INTO [exampleoutput]
        FROM [exampleinput]
                """)
                .build());
    
            var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()
                .name("exampleinput")
                .streamAnalyticsJobName(exampleJob.name())
                .resourceGroupName(exampleJob.resourceGroupName())
                .storageAccountName(exampleAccount.name())
                .storageAccountKey(exampleAccount.primaryAccessKey())
                .storageContainerName(exampleContainer.name())
                .pathPattern("")
                .dateFormat("yyyy/MM/dd")
                .timeFormat("HH")
                .serialization(StreamInputBlobSerializationArgs.builder()
                    .type("Csv")
                    .encoding("UTF8")
                    .fieldDelimiter(",")
                    .build())
                .build());
    
            var exampleOutputBlob = new OutputBlob("exampleOutputBlob", OutputBlobArgs.builder()
                .name("exampleoutput")
                .streamAnalyticsJobName(exampleJob.name())
                .resourceGroupName(exampleJob.resourceGroupName())
                .storageAccountName(exampleAccount.name())
                .storageAccountKey(exampleAccount.primaryAccessKey())
                .storageContainerName(exampleContainer.name())
                .pathPattern("example-{date}-{time}")
                .dateFormat("yyyy-MM-dd")
                .timeFormat("HH")
                .serialization(OutputBlobSerializationArgs.builder()
                    .type("Avro")
                    .build())
                .build());
    
            var exampleJobSchedule = new JobSchedule("exampleJobSchedule", JobScheduleArgs.builder()
                .streamAnalyticsJobId(exampleJob.id())
                .startMode("CustomTime")
                .startTime("2022-09-21T00:00:00Z")
                .build(), CustomResourceOptions.builder()
                    .dependsOn(                
                        exampleJob,
                        exampleStreamInputBlob,
                        exampleOutputBlob)
                    .build());
    
        }
    }
    
    resources:
      example:
        type: azure:core:ResourceGroup
        properties:
          name: example-resources
          location: West Europe
      exampleAccount:
        type: azure:storage:Account
        name: example
        properties:
          name: example
          resourceGroupName: ${example.name}
          location: ${example.location}
          accountTier: Standard
          accountReplicationType: LRS
      exampleContainer:
        type: azure:storage:Container
        name: example
        properties:
          name: example
          storageAccountName: ${exampleAccount.name}
          containerAccessType: private
      exampleBlob:
        type: azure:storage:Blob
        name: example
        properties:
          name: example
          storageAccountName: ${exampleAccount.name}
          storageContainerName: ${exampleContainer.name}
          type: Block
          source:
            fn::FileAsset: example.csv
      exampleJob:
        type: azure:streamanalytics:Job
        name: example
        properties:
          name: example-job
          resourceGroupName: ${example.name}
          location: ${example.location}
          compatibilityLevel: '1.2'
          dataLocale: en-GB
          eventsLateArrivalMaxDelayInSeconds: 60
          eventsOutOfOrderMaxDelayInSeconds: 50
          eventsOutOfOrderPolicy: Adjust
          outputErrorPolicy: Drop
          streamingUnits: 3
          tags:
            environment: Example
          transformationQuery: |2
                SELECT *
                INTO [exampleoutput]
                FROM [exampleinput]
      exampleStreamInputBlob:
        type: azure:streamanalytics:StreamInputBlob
        name: example
        properties:
          name: exampleinput
          streamAnalyticsJobName: ${exampleJob.name}
          resourceGroupName: ${exampleJob.resourceGroupName}
          storageAccountName: ${exampleAccount.name}
          storageAccountKey: ${exampleAccount.primaryAccessKey}
          storageContainerName: ${exampleContainer.name}
          pathPattern:
          dateFormat: yyyy/MM/dd
          timeFormat: HH
          serialization:
            type: Csv
            encoding: UTF8
            fieldDelimiter: ','
      exampleOutputBlob:
        type: azure:streamanalytics:OutputBlob
        name: example
        properties:
          name: exampleoutput
          streamAnalyticsJobName: ${exampleJob.name}
          resourceGroupName: ${exampleJob.resourceGroupName}
          storageAccountName: ${exampleAccount.name}
          storageAccountKey: ${exampleAccount.primaryAccessKey}
          storageContainerName: ${exampleContainer.name}
          pathPattern: example-{date}-{time}
          dateFormat: yyyy-MM-dd
          timeFormat: HH
          serialization:
            type: Avro
      exampleJobSchedule:
        type: azure:streamanalytics:JobSchedule
        name: example
        properties:
          streamAnalyticsJobId: ${exampleJob.id}
          startMode: CustomTime
          startTime: 2022-09-21T00:00:00Z
        options:
          dependson:
            - ${exampleJob}
            - ${exampleStreamInputBlob}
            - ${exampleOutputBlob}
    

    Create JobSchedule Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new JobSchedule(name: string, args: JobScheduleArgs, opts?: CustomResourceOptions);
    @overload
    def JobSchedule(resource_name: str,
                    args: JobScheduleArgs,
                    opts: Optional[ResourceOptions] = None)
    
    @overload
    def JobSchedule(resource_name: str,
                    opts: Optional[ResourceOptions] = None,
                    start_mode: Optional[str] = None,
                    stream_analytics_job_id: Optional[str] = None,
                    start_time: Optional[str] = None)
    func NewJobSchedule(ctx *Context, name string, args JobScheduleArgs, opts ...ResourceOption) (*JobSchedule, error)
    public JobSchedule(string name, JobScheduleArgs args, CustomResourceOptions? opts = null)
    public JobSchedule(String name, JobScheduleArgs args)
    public JobSchedule(String name, JobScheduleArgs args, CustomResourceOptions options)
    
    type: azure:streamanalytics:JobSchedule
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args JobScheduleArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args JobScheduleArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args JobScheduleArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args JobScheduleArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args JobScheduleArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    Constructor example

    The following reference example uses placeholder values for all input properties.

    var azureJobScheduleResource = new Azure.StreamAnalytics.JobSchedule("azureJobScheduleResource", new()
    {
        StartMode = "string",
        StreamAnalyticsJobId = "string",
        StartTime = "string",
    });
    
    example, err := streamanalytics.NewJobSchedule(ctx, "azureJobScheduleResource", &streamanalytics.JobScheduleArgs{
    	StartMode:            pulumi.String("string"),
    	StreamAnalyticsJobId: pulumi.String("string"),
    	StartTime:            pulumi.String("string"),
    })
    
    var azureJobScheduleResource = new JobSchedule("azureJobScheduleResource", JobScheduleArgs.builder()
        .startMode("string")
        .streamAnalyticsJobId("string")
        .startTime("string")
        .build());
    
    azure_job_schedule_resource = azure.streamanalytics.JobSchedule("azureJobScheduleResource",
        start_mode="string",
        stream_analytics_job_id="string",
        start_time="string")
    
    const azureJobScheduleResource = new azure.streamanalytics.JobSchedule("azureJobScheduleResource", {
        startMode: "string",
        streamAnalyticsJobId: "string",
        startTime: "string",
    });
    
    type: azure:streamanalytics:JobSchedule
    properties:
        startMode: string
        startTime: string
        streamAnalyticsJobId: string
    

    JobSchedule Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    The JobSchedule resource accepts the following input properties:

    StartMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    StreamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    StartTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    StartMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    StreamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    StartTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    startMode String

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    streamAnalyticsJobId String
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    startTime String
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    startMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    streamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    startTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    start_mode str

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    stream_analytics_job_id str
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    start_time str
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    startMode String

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    streamAnalyticsJobId String
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    startTime String
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime

    Outputs

    All input properties are implicitly available as output properties. Additionally, the JobSchedule resource produces the following output properties:

    Id string
    The provider-assigned unique ID for this managed resource.
    LastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    Id string
    The provider-assigned unique ID for this managed resource.
    LastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    id String
    The provider-assigned unique ID for this managed resource.
    lastOutputTime String
    The time at which the Stream Analytics job last produced an output.
    id string
    The provider-assigned unique ID for this managed resource.
    lastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    id str
    The provider-assigned unique ID for this managed resource.
    last_output_time str
    The time at which the Stream Analytics job last produced an output.
    id String
    The provider-assigned unique ID for this managed resource.
    lastOutputTime String
    The time at which the Stream Analytics job last produced an output.

    Look up Existing JobSchedule Resource

    Get an existing JobSchedule resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: JobScheduleState, opts?: CustomResourceOptions): JobSchedule
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            last_output_time: Optional[str] = None,
            start_mode: Optional[str] = None,
            start_time: Optional[str] = None,
            stream_analytics_job_id: Optional[str] = None) -> JobSchedule
    func GetJobSchedule(ctx *Context, name string, id IDInput, state *JobScheduleState, opts ...ResourceOption) (*JobSchedule, error)
    public static JobSchedule Get(string name, Input<string> id, JobScheduleState? state, CustomResourceOptions? opts = null)
    public static JobSchedule get(String name, Output<String> id, JobScheduleState state, CustomResourceOptions options)
    Resource lookup is not supported in YAML
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    LastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    StartMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    StartTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    StreamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    LastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    StartMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    StartTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    StreamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    lastOutputTime String
    The time at which the Stream Analytics job last produced an output.
    startMode String

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    startTime String
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    streamAnalyticsJobId String
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    lastOutputTime string
    The time at which the Stream Analytics job last produced an output.
    startMode string

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    startTime string
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    streamAnalyticsJobId string
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    last_output_time str
    The time at which the Stream Analytics job last produced an output.
    start_mode str

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    start_time str
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    stream_analytics_job_id str
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
    lastOutputTime String
    The time at which the Stream Analytics job last produced an output.
    startMode String

    The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

    Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

    startTime String
    The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
    streamAnalyticsJobId String
    The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.

    Import

    Stream Analytics Job’s can be imported using the resource id, e.g.

    $ pulumi import azure:streamanalytics/jobSchedule:JobSchedule example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/schedule/default
    

    To learn more about importing existing cloud resources, see Importing resources.

    Package Details

    Repository
    Azure Classic pulumi/pulumi-azure
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the azurerm Terraform Provider.
    azure logo

    We recommend using Azure Native.

    Azure Classic v5.81.0 published on Monday, Jun 24, 2024 by Pulumi