section> Computing
  • Auto Scaling
  • Bare Metal Server
  • Dedicated Host
  • Elastic Cloud Server
  • FunctionGraph
  • Image Management Service
Network
  • Direct Connect
  • Domain Name Service
  • Elastic IP
  • Elastic Load Balancing
  • Enterprise Router
  • NAT Gateway
  • Private Link Access Service
  • Secure Mail Gateway
  • Virtual Private Cloud
  • VPC Endpoint
  • Virtual Private Network
Storage
  • Cloud Backup and Recovery
  • Cloud Server Backup Service
  • Elastic Volume Service
  • Object Storage Service
  • Storage Disaster Recovery Service
  • Scalable File Service
  • Volume Backup Service
Application Services
  • Application Operations Management
  • Application Performance Management
  • API Gateway (APIG)
  • Distributed Message Service
  • Simple Message Notification
Database Services
  • Distributed Cache Service
  • Document Database Service
  • Data Replication Service
  • GaussDB (for MySQL)
  • GeminiDB
  • Relational Database Service
Big Data and Data Analysis
  • Cloud Search Service
  • DataArts Studio
  • Data Ingestion Service
  • Data Lake Insight
  • Data Warehouse Service
  • ModelArts
  • MapReduce Service
  • Optical Character Recognition
Container Services
  • Cloud Container Engine
  • Software Repository for Containers
Applications and Databases
  • Anti DDoS
  • Database Security Service
  • Host Security Service
  • Web Application Firewall
  • Dedicated Web Application Firewall
Identity and Access Management
  • Identity and Access Management Service
Key Management
  • Key Management Service
Compliance
  • Core Services Certifications
Monitoring and Logging
  • Cloud Eye
  • Cloud Trace Service
  • Log Tank Service
Resource Management
  • Cloud Create
  • Resource Management Service
  • Tag Management Service
Other
  • Enterprise Dashboard
  • Price API
  • Price Calculator
APIs
  • REST API
  • API Usage Guidelines
  • Endpoints
Development and Automation
  • SDKs
  • Drivers and Tools
  • Terraform
  • Ansible
  • Cloud Create
Architecture Center
  • Best Practices
  • Blueprints
Other
  • Status Dashboard
  • Portfolio Roadmap
  • Training Certifications
IaaSComputingAuto ScalingBare Metal ServerDedicated HostElastic Cloud ServerFunctionGraphImage Management ServiceNetworkDirect ConnectDomain Name ServiceElastic IPElastic Load BalancingEnterprise RouterNAT GatewayPrivate Link Access ServiceSecure Mail GatewayVirtual Private CloudVPC EndpointVirtual Private NetworkStorageCloud Backup and RecoveryCloud Server Backup ServiceElastic Volume ServiceObject Storage ServiceStorage Disaster Recovery ServiceScalable File ServiceVolume Backup ServicePaaSApplications ServicesApplication Operation ManagementApplication Performance ManagementAPI Gateway (APIG)Distributed Message ServiceSimple Message NotificationDatabase ServicesDistributed Cache ServiceDocument Database ServiceData Replication ServiceGaussDB (for MySQL)GeminiDBRelational Database ServiceBig Data and Data AnalysisCloud Search ServiceDataArts StudioData Ingestion ServiceData Lake InsightData Warehouse ServiceModelArtsMapReduce ServiceOptical Character RecognitionContainer ServicesCloud Container EngineSoftware Repository for ContainersSecurityApplications and DatabasesAnti DDoSDatabase Security ServiceHost Security ServiceWeb Application FirewallDedicated Web Application FirewallIdentity and Access ManagementIdentity and Access Management ServiceKey ManagementKey Management ServiceComplianceCore Services CertificationsManagementMonitoring and LoggingCloud EyeCloud Tracking ServiceLog Tank ServiceResource ManagementCloud CreateResource Management ServiceTag Management ServiceOtherEnterprise DashboardPrice APIPrice CalculatorDevelopersAPIsREST APIAPI Usage GuidelinesEndpointsAutomation and DevelopmentSDKsDrivers and ToolsTerraformAnsibleCloud CreateArchitecture CenterBest PracticesBlueprintsOtherStatus DashboardPortfolio RoadmapTraining Certifications

Data Lake Insight

  • Service Overview
  • Getting Started
  • DLI Console Overview
  • SQL Editor
  • Job Management
    • Overview
    • SQL Job Management
    • Flink Job Management
      • Overview
      • Managing Flink Job Permissions
      • Preparing Flink Job Data
      • (Recommended) Creating a Flink OpenSource SQL Job
      • Creating a Flink SQL Job
      • Creating a Flink Jar Job
      • Performing Operations on a Flink Job
      • Flink Job Details
      • Tag Management
      • Enabling Dynamic Scaling for Flink Jobs
    • Spark Job Management
    • Setting the Priority for a Job
  • Queue Management
  • Elastic Resource Pool
  • Data Management
  • Job Templates
  • Enhanced Datasource Connections
  • Datasource Authentication
  • Global Configuration
  • Permissions Management
  • Other Common Operations
  • FAQ
  • Change History
  • User Guide
  • Job Management
  • Flink Job Management
  • Performing Operations on a Flink Job

Performing Operations on a Flink Job¶

After a job is created, you can perform operations on the job as required.

  • Editing a Job

  • Starting a Job

  • Stopping a Job

  • Deleting a Job

  • Exporting a Job

  • Importing a Job

  • Modifying Name and Description

  • Importing to a Savepoint

  • Triggering a Savepoint

  • Runtime Configuration

Editing a Job¶

You can edit a created job, for example, by modifying the SQL statement, job name, job description, or job configurations.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  2. In the row where the job you want to edit locates, click Edit in the Operation column to switch to the editing page.

  3. Edit the job as required.

    For details about how to edit a Flink SQL job, see Step 5 to Step 7 in Creating a Flink SQL Job.

    For details about how to edit a user-defined Flink job, see Step 5 to Step 7 in Creating a Flink Jar Job.

Starting a Job¶

You can start a saved or stopped job.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  2. Use either of the following methods to start jobs:

    • Starting a single job

      Select a job and click Start in the Operation column.

      Alternatively, you can select the row where the job you want to start locates and click Start in the upper left of the job list.

    • Batch starting jobs

      Select the rows where the jobs you want to start locate and click Start in the upper left of the job list.

    After you click Start, the Start Flink Jobs page is displayed.

  1. On the Start Flink Jobs page, confirm the job information. If they are correct, click Start Now.

    After a job is started, you can view the job execution result in the Status column.

Stopping a Job¶

You can stop a job in the Running or Submitting state.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  2. Stop a job using either of the following methods:

    • Stopping a job

      Locate the row that contains the job to be stopped, click More in the Operation column, and select Stop.

      Alternatively, you can select the row where the job you want to stop locates and click Stop in the upper left of the job list.

    • Batch stopping jobs

      Locate the rows containing the jobs you want to stop and click Stop in the upper left of the job list.

  3. In the displayed Stop Job dialog box, click OK to stop the job.

    Note

    • Before stopping a job, you can trigger a savepoint to save the job status information. When you start the job again, you can choose whether to restore the job from the savepoint.

    • If you select Trigger savepoint, a savepoint is created. If Trigger savepoint is not selected, no savepoint is created. By default, the savepoint function is disabled.

    • The lifecycle of a savepoint starts when the savepoint is triggered and stops the job, and ends when the job is restarted. The savepoint is automatically deleted after the job is restarted.

    When a job is being stopped, the job status is displayed in the Status column of the job list. The details are as follows:

    • Stopping: indicates that the job is being stopped.

    • Stopped: indicates that the job is stopped successfully.

    • Stop failed: indicates that the job failed to be stopped.

Deleting a Job¶

If you do not need to use a job, perform the following operations to delete it. A deleted job cannot be restored. Therefore, exercise caution when deleting a job.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  1. Perform either of the following methods to delete jobs:

    • Deleting a single job

      Locate the row containing the job you want to delete and click More > Delete in the Operation column.

      Alternatively, you can select the row containing the job you want to delete and click Delete in the upper left of the job list.

    • Deleting jobs in batches

      Select the rows containing the jobs you want to delete and click Delete in the upper left of the job list.

  2. Click Yes.

Exporting a Job¶

You can export the created Flink jobs to an OBS bucket.

This mode is applicable to the scenario where a large number of jobs need to be created when you switch to another region, project, or user. In this case, you do not need to create a job. You only need to export the original job, log in to the system in a new region or project, or use a new user to import the job.

Note

When switching to another project or user, you need to grant permissions to the new project or user. For details, see Managing Flink Job Permissions.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  1. Click Export Job in the upper right corner. The Export Job dialog box is displayed.

  2. Select the OBS bucket where the job is stored. Click Next.

  3. Select job information you want to export.

    By default, configurations of all jobs are exported. You can enable the Custom Export function to export configurations of the desired jobs.

  4. Click Confirm to export the job.

Importing a Job¶

You can import the Flink job configuration file stored in the OBS bucket to the Flink Jobs page of DLI.

This mode is applicable to the scenario where a large number of jobs need to be created when you switch to another region, project, or user. In this case, you do not need to create a job. You only need to export the original job, log in to the system in a new region or project, or use a new user to import the job.

If you need to import a self-created job, you are advised to use the job creation function. For details, see Creating a Flink SQL Job and Creating a Flink Jar Job.

Note

  • When switching to another project or user, you need to grant permissions to the new project or user. For details, see Managing Flink Job Permissions.

  • Only jobs whose data format is the same as that of Flink jobs exported from DLI can be imported.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  1. Click Import Job in the upper right corner. The Import Job dialog box is displayed.

  2. Select the complete OBS path of the job configuration file to be imported. Click Next.

  3. Configure the same-name job policy and click next. Click Next.

    • Select Overwrite job of the same name. If the name of the job to be imported already exists, the existing job configuration will be overwritten and the job status switches to Draft.

    • If Overwrite job of the same name is not selected and the name of the job to be imported already exists, the job will not be imported.

  4. Ensure that Config File and Overwrite Same-Name Job are correctly configured. Click Confirm to import the job.

Modifying Name and Description¶

You can change the job name and description as required.

  1. In the left navigation pane of the DLI management console, choose Job Management > Flink Jobs. The Flink Jobs page is displayed.

  2. In the Operation column of the job whose name and description need to be modified, choose More > Modify Name and Description. The Modify Name and Description dialog box is displayed. Change the name or modify the description of a job.

  3. Click OK.

Triggering a Savepoint¶

When you need to stop a job, you can create a savepoint to save the job status information. In this case, when you restart the job, you can choose to restore the job from the latest savepoint.

Note

  • You can click Trigger Savepoint for jobs in the Running status to save the job status.

  • The lifecycle of a savepoint starts when the savepoint is triggered and stops the job, and ends when the job is restarted. The savepoint is automatically deleted after the job is restarted.

Importing to a Savepoint¶

You can import a savepoint to restore the job status. For details about the savepoint, see Checkpointing at the official website of Flink.

You need to select the OBS path of the save point.

Runtime Configuration¶

You can select Runtime Configuration to configure job exception alarms and restart options.

Note

Flink SQL jobs and Flink Jar jobs are supported.

  1. In the Operation column of the Flink job, choose More > Runtime Configuration.

  2. In the Runtime Configuration dialog box, set the following parameters:

    Table 1 Running parameters¶

    Parameter

    Description

    Name

    Job name.

    Alarm Generation upon Job Exception

    Whether to report job exceptions, for example, abnormal job running or exceptions due to an insufficient balance, to users via SMS or email.

    If this option is selected, you need to set the following parameters:

    SMN Topic

    Select a user-defined SMN topic. For details about how to customize SMN topics, see Creating a Topic in the Simple Message Notification User Guide.

    Auto Restart upon Exception

    Whether to enable automatic restart. If this function is enabled, any job that has become abnormal will be automatically restarted.

    If this option is selected, you need to set the following parameters:

    • Max. Retry Attempts: maximum number of retry times upon an exception. The unit is times/hour.

      • Unlimited: The number of retries is unlimited.

      • Limited: The number of retries is user-defined.

    • Restore Job from Checkpoint: Restore the job from the saved checkpoint.

      Note

      This parameter cannot be configured for Flink SQL jobs or Flink OpenSource SQL jobs.

      If this parameter is selected, you need to set Checkpoint Path for Flink Jar jobs.

      Checkpoint Path: Select the checkpoint saving path. The checkpoint path must be the same as that you set in the application package. Note that the checkpoint path for each job must be unique. Otherwise, the checkpoint cannot be obtained.

  • Prev
  • Next
last updated: 2025-04-11 13:47 UTC - commit: aa96e2e0eed630617f88dab58364d89a112d472d
Edit pageReport Documentation Bug
Page Contents
  • Performing Operations on a Flink Job
    • Editing a Job
    • Starting a Job
    • Stopping a Job
    • Deleting a Job
    • Exporting a Job
    • Importing a Job
    • Modifying Name and Description
    • Triggering a Savepoint
    • Importing to a Savepoint
    • Runtime Configuration
© T-Systems International GmbH
  • Contact
  • Data privacy
  • Disclaimer of liabilitys
  • Imprint