flink operator 1.7 更换日志框架log4j 到logback

更换日志框架 flink 1.18

1 消除基础flink框架log4j 添加logback jar

bash 复制代码
1-1 log4j
log4j-1.2-api-2.17.1.jar
log4j-api-2.17.1.jar
log4j-core-2.17.1.jar
log4j-slf4j-impl-2.17.1.jar
1-2  logback
logback-core-1.2.3.jar
logback-classic-1.2.3.jar
slf4j-api-1.7.25.jar

2 在镜像制作的时候消除:

bash 复制代码
RUN rm -rf $FLINK_HOME/lib/log4j-1.2-api-2.17.1.jar
RUN rm -rf $FLINK_HOME/lib/log4j-api-2.17.1.jar
RUN rm -rf $FLINK_HOME/lib/log4j-core-2.17.1.jar
RUN rm -rf $FLINK_HOME/lib/log4j-slf4j-impl-2.17.1.jar
RUN rm -rf $FLINK_HOME/lib/log4j-1.2-api-2.17.1.jar
COPY logback-core-1.2.3.jar $FLINK_HOME/lib
COPY logback-classic-1.2.3.jar $FLINK_HOME/lib
COPY slf4j-api-1.7.25.jar $FLINK_HOME/lib      

3-2 制作flink yaml的时候添加参数

yaml 复制代码
logConfiguration:
    logback-console.xml: '<configuration>    <appender name="file" class="ch.qos.logback.core.rolling.RollingFileAppender">        <file>${log.file}</file>        <Prudent>false</Prudent>        <encoder>            <pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %-5level %logger{60} %X{sourceThread} - %msg%n</pattern>        </encoder>        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">            <fileNamePattern>${log.file}.%d{yyyy-MM-dd.HH.mm}</fileNamePattern>            <MaxHistory>15</MaxHistory>        </rollingPolicy>    </appender>    <appender name="localfile" class="ch.qos.logback.core.rolling.RollingFileAppender">        <file>/opt/flink/log/${log.file}</file>        <append>false</append>        <Prudent>true</Prudent>        <encoder>            <pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %-5level %logger{60} %X{sourceThread} - %msg%n</pattern>        </encoder>        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">            <fileNamePattern>/opt/flink/log/${log.file}.%d{yyyy-MM-dd.HH.mm}</fileNamePattern>            <MaxHistory>15</MaxHistory>        </rollingPolicy>    </appender>    <root level="DEBUG">        <appender-ref ref="file"/>    </root>    <logger name="akka" level="INFO">        <appender-ref ref="file"/>    </logger>    <logger name="org.apache.kafka" level="INFO">        <appender-ref ref="file"/>    </logger>    <logger name="org.apache.hadoop" level="INFO">        <appender-ref ref="file"/>    </logger>    <logger name="org.apache.zookeeper" level="INFO">        <appender-ref ref="file"/>    </logger>    <logger name="org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline" level="ERROR">        <appender-ref ref="file"/>    </logger></configuration>'
相关推荐
w***765513 小时前
SpringBoot Test详解
spring boot·后端·log4j
Jackyzhe14 小时前
Flink源码阅读:Mailbox线程模型
大数据·flink
Hello.Reader16 小时前
Flink 读文本文件TextLineInputFormat + FileSource(批/流一体)+ 目录持续监控
大数据·flink
Analyze_ing16 小时前
DolphinScheduler启动flink任务, 用Flink消费Kafka数据(linux)
大数据·flink·kafka
面向Google编程1 天前
Flink源码阅读:Mailbox线程模型
大数据·flink
Hello.Reader2 天前
CSV Format Flink / PyFlink 读写 CSV 的正确姿势(含 Schema 高级配置)
大数据·python·flink
Knight_AL2 天前
Maven 生命周期详解(validate → deploy)
java·log4j·maven
Hello.Reader2 天前
Hadoop Formats 在 Flink 里复用 Hadoop InputFormat(flink-hadoop-compatibility)
大数据·hadoop·flink
岁岁种桃花儿2 天前
Spring Boot核心插件全解析(官方+第三方,附使用场景)
log4j·springboot·插件
Jackeyzhe2 天前
Flink源码阅读:JobManager的HA机制
flink