section> Computing
  • Auto Scaling
  • Bare Metal Server
  • Dedicated Host
  • Elastic Cloud Server
  • FunctionGraph
  • Image Management Service
Network
  • Direct Connect
  • Domain Name Service
  • Elastic IP
  • Elastic Load Balancing
  • Enterprise Router
  • NAT Gateway
  • Private Link Access Service
  • Secure Mail Gateway
  • Virtual Private Cloud
  • Virtual Private Network
  • VPC Endpoint
Storage
  • Cloud Backup and Recovery
  • Cloud Server Backup Service
  • Elastic Volume Service
  • Object Storage Service
  • Scalable File Service
  • Storage Disaster Recovery Service
  • Volume Backup Service
Application
  • API Gateway (APIG)
  • Application Operations Management
  • Application Performance Management
  • Distributed Message Service (for Kafka)
  • Simple Message Notification
Data Analysis
  • Cloud Search Service
  • Data Lake Insight
  • Data Warehouse Service
  • DataArts Studio
  • MapReduce Service
  • ModelArts
  • Optical Character Recognition
Container
  • Application Service Mesh
  • Cloud Container Engine
  • Cloud Container Instance
  • Software Repository for Containers
Databases
  • Data Replication Service
  • Distributed Cache Service
  • Distributed Database Middleware
  • Document Database Service
  • GaussDB (for MySQL)
  • GeminiDB
  • Relational Database Service
Management & Deployment
  • Cloud Create
  • Cloud Eye
  • Cloud Trace Service
  • Config
  • Log Tank Service
  • Resource Formation Service
  • Tag Management Service
Security Services
  • Anti-DDoS
  • Cloud Firewall
  • Database Security Service
  • Dedicated Web Application Firewall
  • Host Security Service
  • Identity and Access Management
  • Key Management Service
  • Web Application Firewall
Other
  • Enterprise Dashboard
  • Marketplace
  • Price Calculator
  • Status Dashboard
APIs
  • REST API
  • API Usage Guidelines
  • Endpoints
Development and Automation
  • SDKs
  • Drivers and Tools
  • Terraform
  • Ansible
  • Cloud Create
Architecture Center
  • Best Practices
  • Blueprints
IaaSComputingAuto ScalingBare Metal ServerDedicated HostElastic Cloud ServerFunctionGraphImage Management ServiceNetworkDirect ConnectDomain Name ServiceElastic IPElastic Load BalancingEnterprise RouterNAT GatewayPrivate Link Access ServiceSecure Mail GatewayVirtual Private CloudVirtual Private NetworkVPC EndpointStorageCloud Backup and RecoveryCloud Server Backup ServiceElastic Volume ServiceObject Storage ServiceScalable File ServiceStorage Disaster Recovery ServiceVolume Backup ServicePaaSApplicationAPI Gateway (APIG)Application Operations ManagementApplication Performance ManagementDistributed Message Service (for Kafka)Simple Message NotificationData AnalysisCloud Search ServiceData Lake InsightData Warehouse ServiceDataArts StudioMapReduce ServiceModelArtsOptical Character RecognitionContainerApplication Service MeshCloud Container EngineCloud Container InstanceSoftware Repository for ContainersDatabasesData Replication ServiceDistributed Cache ServiceDistributed Database MiddlewareDocument Database ServiceGaussDB (for MySQL)GeminiDBRelational Database ServiceManagementManagement & DeploymentCloud CreateCloud EyeCloud Trace ServiceConfigLog Tank ServiceResource Formation ServiceTag Management ServiceSecuritySecurity ServicesAnti-DDoSCloud FirewallDatabase Security ServiceDedicated Web Application FirewallHost Security ServiceIdentity and Access ManagementKey Management ServiceWeb Application FirewallOtherOtherEnterprise DashboardMarketplacePrice CalculatorStatus Dashboard

DataArts Studio

  • Service Overview
  • Preparations
  • User Guide
    • Preparations Before Using DataArts Studio
    • Management Center
    • DataArts Migration
    • DataArts Factory
      • Overview
      • Data Management
      • Script Development
      • Job Development
      • Solution
      • Execution History
      • O&M and Scheduling
      • Configuration and Management
      • Node Reference
      • EL Expression Reference
      • Usage Guidance
        • Job Dependency
        • IF Statements
        • Obtaining the Return Value of a Rest Client Node
        • Using For Each Nodes
        • Developing a Python Script
        • Developing a DWS SQL Job
        • Developing a Hive SQL Job
        • Developing a DLI Spark Job
        • Developing an MRS Flink Job
        • Developing an MRS Spark Python Job
  • FAQs
  • Change History
  • User Guide
  • User Guide
  • DataArts Factory
  • Usage Guidance
  • Developing an MRS Flink Job

Developing an MRS Flink Job¶

This section describes how to develop an MRS Flink job on DataArts Factory. Use an MRS Flink job to count the number of words.

Prerequisites¶

  • You have the permission to access OBS paths.

  • MRS has been enabled and an MRS cluster has been created.

Data Preparation¶

  • Download the Flink job resource package wordcount.jar from https://github.com/apache/flink/tree/master/flink-examples/flink-examples-streaming/src/main/java/org/apache/flink/streaming/examples/wordcount.

  • Prepare the data file in.txt, which contains some English words.

Procedure¶

  1. Upload the job resource package and data file to the OBS bucket.

    Note

    In this example, upload WordCount.jar to lkj_test/WordCount.jar and word.txt to lkj_test/input/word.txt.

  2. Create an empty job named job_MRS_Flink.

    **Figure 1** Creating a job

    Figure 1 Creating a job¶

  3. Go to the job development page, drag the MRS Flink node to the canvas, and click the node to configure its properties.

    **Figure 2** Configuring properties for an MRS Flink node

    Figure 2 Configuring properties for an MRS Flink node¶

    Parameter descriptions:

    --Flink job name
    wordcount
    --MRS cluster name
    Select an MRS cluster.
    --Program parameter
    -c    org.apache.flink.streaming.examples.wordcount.WordCount
    --Flink job resource package
    wordcount
    --Input data path
    obs://dlf-beijing2/lkj_test/input/word.txt
    --Output data path
    obs://dlf-beijing2/lkj_test/output.txt
    

    Specifically:

    obs://dlf-beijing2/lkj_test/input/word.txt is the directory where the wordcount.jar parameters are passed. You can pass the words to count.

    obs://dlf-beijing2/lkj_test/output.txt is the directory where the output parameter file is stored. (If the output.txt file already exists, an error is reported.)

  4. Click Test to execute the MRS Flink job.

  5. After the test is complete, click Submit.

  6. Choose Monitor Job in the navigation pane and view the job execution result.

  7. View the returned records in the OBS bucket. (Skip this step if the return function is not configured.)

  • Prev
  • Next
last updated: 2025-07-09 14:35 UTC - commit: d615b325251d87083330e48e30efad3971e0fa88
Edit pageReport Documentation Bug
Page Contents
  • Developing an MRS Flink Job
    • Prerequisites
    • Data Preparation
    • Procedure
© T-Systems International GmbH
  • Contact
  • Data privacy
  • Disclaimer of liabilitys
  • Imprint