在k8s部署flink standaloneSession模式 job-manager 高可用集群搭建

1. 创建namespace及serviceaccount
创建kubernets命名空间
  kubectl create namespace app-flink
创建该空间的用户 flink (自定义)
  kubectl create serviceaccount flink -n  app-flink
给该用户赋权
kubectl create clusterrolebinding flink-role-binding-flink \
        --clusterrole=edit \
        --serviceaccount=app-flink:flink

(不创建用户和赋权启动时候会报 "权限"错)
kubectl create clusterrolebinding endpoints-reader-mydefault \
  --clusterrole=endpoints-reader  \
  --serviceaccount=app-flink:flink
2. 在k8s上部署minio

① 创建minio-pvc.yml

apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: pvc-minio
  namespace: app-flink
spec:
  accessModes:
    - ReadWriteOnce
  resources:
    requests:
      storage: 40Gi

执行命令:kubectl create -f minio-pvc.yml

② 创建minio的deployment及service(minio.yml)
# 添加
apiVersion: v1
kind: Service
metadata:
  name: minio-svc
  namespace: app-flink
  labels:
    app: minio
spec:
  type: NodePort
  ports:
    - name: minio-port
      protocol: TCP
      port: 9000
      nodePort: 32600
      targetPort: 9000
    - name: minio-http
        protocol: TCP
        port: 9009
        nodePort: 32601
        targetPort: 9009
  selector:
    app: minio
---
apiVersion: v1
kind: Pod
metadata:
  name: minio
  namespace: app-flink
  labels:
    app: minio
spec:
  containers:
    - name: minio
      env:
        - name: MINIO_ACCESS_KEY
          value: "admin"
        - name: MINIO_SECRET_KEY
          value: "admin123"
      image: minio/minio:latest
      args: ["server", "/data/minio", "--console-address=0.0.0.0:9009"]
      ports:
        - name: minio
          containerPort: 9000
        - name: http
          containerPort: 9009
      volumeMounts:
        - name: data
          mountPath: /data/minio
  volumes:
    - name: data
      persistentVolumeClaim:
        claimName: pvc-minio

在对应的节点上创建目录:mkdir -p /data/minio

执行命令:kubectl create -f minio.yml

3.创建flink配置文件(flink-configuration-configmap.yml)
apiVersion: v1
kind: ConfigMap
metadata:
  name: flink-session-config
  namespace: app-flink
  labels:
    app: flink
data:
  flink-conf.yaml: |+
    kubernetes.cluster-id: app-flink-session
    # 所在的命名空间
    kubernetes.namespace: app-flink
    jobmanager.rpc.address: flink-jobmanager
    taskmanager.numberOfTaskSlots: 16
    blob.server.port: 6124
    jobmanager.rpc.port: 6123
    taskmanager.rpc.port: 6122
    queryable-state.proxy.ports: 6125
    jobmanager.memory.process.size: 1600m
    taskmanager.memory.process.size: 2867m
    parallelism.default: 2
    execution.checkpointing.interval: 10s    
    # 文件系统
    fs.default-scheme: s3
    # s3地址
    s3.endpoint: https://app-flink.s3.cn-north-1.amazonaws.com.cn/
    # s3的bucket
    s3.flink.bucket: app-flink
    s3.access-key: AKIA2MCJLQSFKHMFQ47Q
    s3.secret-key: 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
    # 状态存储格式
    state.backend: rocksdb
    s3.path.style.access: true
    blob.storage.directory: /opt/flink/tmp/blob
    web.upload.dir: /opt/flink/tmp/upload
    io.tmp.dirs: /opt/flink/tmp
    # 状态管理
    # checkpoint存储地址
    state.checkpoints.dir: s3://app-flink/state/checkpoint
    # savepoint存储地址
    state.savepoints.dir: s3://app-flink/state/savepoint
    # checkpoint间隔
    execution.checkpointing.interval: 5000
    execution.checkpointing.mode: EXACTLY_ONCE
    # checkpoint保留数量
    state.checkpoints.num-retained: 3
    # history-server# 监视以下目录中已完成的作业
    jobmanager.archive.fs.dir: s3://app-flink/completed-jobs
    # 每 10 秒刷新一次
    historyserver.archive.fs.refresh-interval: 10000
    historyserver.archive.fs.dir: s3://app-flink/completed-jobs
    # 高可用
    high-availability.type: kubernetes
    high-availability: org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory
    high-availability.storageDir: s3://app-flink/ha
    # 每6个小时触发一次savepoint
    kubernetes.operator.periodic.savepoint.interval: 6h
    kubernetes.operator.savepoint.history.max.age: 24h
    kubernetes.operator.savepoint.history.max.count: 5
    # Restart of unhealthy job deployments
    kubernetes.operator.cluster.health-check.enabled: true
    # Restart failed job deployments 
    kubernetes.operator.job.restart.failed: true
  log4j-console.properties: |+
    # This affects logging for both user code and Flink
    rootLogger.level = INFO
    rootLogger.appenderRef.console.ref = ConsoleAppender
    rootLogger.appenderRef.rolling.ref = RollingFileAppender

    # Uncomment this if you want to _only_ change Flink's logging
    #logger.flink.name = org.apache.flink
    #logger.flink.level = INFO

    # The following lines keep the log level of common libraries/connectors on
    # log level INFO. The root logger does not override this. You have to manually
    # change the log levels here.
    logger.akka.name = akka
    logger.akka.level = INFO
    logger.kafka.name= org.apache.kafka
    logger.kafka.level = INFO
    logger.hadoop.name = org.apache.hadoop
    logger.hadoop.level = INFO
    logger.zookeeper.name = org.apache.zookeeper
    logger.zookeeper.level = INFO

    # Log all infos to the console
    appender.console.name = ConsoleAppender
    appender.console.type = CONSOLE
    appender.console.layout.type = PatternLayout
    appender.console.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x - %m%n

    # Log all infos in the given rolling file
    appender.rolling.name = RollingFileAppender
    appender.rolling.type = RollingFile
    appender.rolling.append = false
    appender.rolling.fileName = ${sys:log.file}
    appender.rolling.filePattern = ${sys:log.file}.%i
    appender.rolling.layout.type = PatternLayout
    appender.rolling.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x - %m%n
    appender.rolling.policies.type = Policies
    appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
    appender.rolling.policies.size.size=100MB
    appender.rolling.strategy.type = DefaultRolloverStrategy
    appender.rolling.strategy.max = 10

    # Suppress the irrelevant (wrong) warnings from the Netty channel handler
    logger.netty.name = org.jboss.netty.channel.DefaultChannelPipeline
    logger.netty.level = OF
4.整合flink及flink cdc构建Docker镜像
FROM flink:1.17.2-scala_2.12

# 需要用到的jar包
# flink-cdc
ADD flink-sql-connector-mysql-cdc-2.4.0.jar /opt/flink/lib/
# jdbc连接器
ADD flink-connector-jdbc-3.1.1-1.17.jar /opt/flink/lib/
# mysql驱动
ADD mysql-connector-j-8.0.31.jar /opt/flink/lib/

RUN mkdir /opt/flink/plugins/s3-fs-presto && cp -f /opt/flink/opt/flink-s3-fs-presto-1.17.2.jar /opt/flink/plugins/s3-fs-presto/

Docker镜像编译命令 docker build -t 10.10.0.203:8188/app/flink:1.17.2_scala_12 .

5.创建jobmanager-pvc.yml
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: flink-tmp
  namespace: app-flink
spec:
  accessModes:
    - ReadWriteOnce
  resources:
    requests:
      storage: 40Gi

执行命令:kubectl create -f jobmanager-pvc.yml

6.创建jobmanager-deployment.yml 
apiVersion: apps/v1
kind: Deployment
metadata:
  name: flink-jobmanager
  namespace: app-flink
spec:
  replicas: 2 # Set the value to greater than 1 to start standby JobManagers
  selector:
    matchLabels:
      app: flink
      component: jobmanager
  template:
    metadata:
      labels:
        app: flink
        component: jobmanager
    spec:
      containers:
        - name: jobmanager
          imagePullPolicy: Always
          image: 10.10.0.203:8188/app/flink:1.17.2_scala_12
          env:
            # 注入POD的ip到容器内
            - name: POD_IP
              valueFrom:
                fieldRef:
                  apiVersion: v1
                  fieldPath: status.podIP
            # 时区
            - name: TZ
              value: Asia/Shanghai
          # The following args overwrite the value of jobmanager.rpc.address configured in the configuration config map to POD_IP.
          args: ["jobmanager", "$(POD_IP)"]
          ports:
            - containerPort: 6123
              name: rpc
            - containerPort: 6124
              name: blob-server
            - containerPort: 8081
              name: webui
          livenessProbe:
            tcpSocket:
              port: 6123
            initialDelaySeconds: 30
            periodSeconds: 60
          resources:
            requests:
              memory: "512Mi"
              cpu: "0.1"
            limits:
              memory: "4096Mi"
              cpu: "2"
          volumeMounts:
            - name: flink-ha-config-volume
              mountPath: /opt/flink/conf
            - name: tmp-dir
              mountPath: /opt/flink/tmp
          securityContext:
            runAsUser: 9999  # refers to user _flink_ from official flink image, change if necessary
      serviceAccountName: flink # Service account which has the permissions to create, edit, delete ConfigMaps
      volumes:
        - name: flink-ha-config-volume
          configMap:
            name: flink-session-config
            items:
              - key: flink-conf.yaml
                path: flink-conf.yaml
              - key: log4j-console.properties
                path: log4j-console.properties
        - name: tmp-dir
          persistentVolumeClaim:
            claimName: flink-tmp
7.创建jobmanager-service-rest.yml

 

apiVersion: v1
kind: Service
metadata:
  name: flink-jobmanager-rest
  namespace: app-flink
spec:
  type: NodePort
  ports:
    - name: rest
      port: 8081
      targetPort: 8081
      nodePort: 30081
  selector:
    app: flink
    component: jobmanager

执行命令:kubectl create -f jobmanager-service-rest.yml

8.创建taskmanager-deployment.yml
apiVersion: apps/v1
kind: Deployment
metadata:
  name: flink-taskmanager
  namespace: app-flink
spec:
  replicas: 2
  selector:
    matchLabels:
      app: flink-taskmanager
      component: taskmanager
  template:
    metadata:
      labels:
        app: flink-taskmanager
        component: taskmanager
    spec:
      containers:
        - name: taskmanager
          image: 10.10.0.203:8188/app/flink:1.17.2_scala_12
          args: ["taskmanager"]
          ports:
            - containerPort: 6122
              name: rpc
          livenessProbe:
            tcpSocket:
              port: 6122
            initialDelaySeconds: 30
            periodSeconds: 60
          volumeMounts:
            - name: flink-ha-config-volume
              mountPath: /opt/flink/conf/
          securityContext:
            runAsUser: 9999  # refers to user _flink_ from official flink image, change if necessary
      serviceAccountName: flink
      volumes:
        - name: flink-ha-config-volume
          configMap:
            name: flink-session-config
            items:
              - key: flink-conf.yaml
                path: flink-conf.yaml
              - key: log4j-console.properties
                path: log4j-console.properties

执行命令:kubectl create -f taskmanager-deployment.yml

  • 7
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

yongge

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值