flink-cdc实战之oracle问题记录01

记录问题,温暖你我,上台

欢迎点赞留言关注

2024-01-26 11:02:56,168 ERROR  Oracle|oracle_logminer|streaming  Mining session stopped due to the {}   [io.debezium.connector.oracle.logminer.LogMinerHelper]
io.debezium.DebeziumException: Supplemental logging not configured for table ORCL.AA.A.  Use command: ALTER TABLE AA.A ADD SUPPLEMENTAL LOG DATA (ALL) COLUMNS
	at io.debezium.connector.oracle.logminer.LogMinerHelper.checkSupplementalLogging(LogMinerHelper.java:407)
	at io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource.execute(LogMinerStreamingChangeEventSource.java:132)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:152)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:119)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:834)
2024-01-26 11:02:56,170 ERROR  Oracle|oracle_logminer|streaming  Producer failure   [io.debezium.pipeline.ErrorHandler]
io.debezium.DebeziumException: Supplemental logging not configured for table ORCL.AA.A.  Use command: ALTER TABLE AA.A ADD SUPPLEMENTAL LOG DATA (ALL) COLUMNS
	at io.debezium.connector.oracle.logminer.LogMinerHelper.checkSupplementalLogging(LogMinerHelper.java:407)
	at io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource.execute(LogMinerStreamingChangeEventSource.java:132)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:152)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:119)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:834)
2024-01-26 11:02:56,170 INFO   Oracle|oracle_logminer|streaming  startScn=3460974, endScn=null, offsetContext.getScn()=3460974   [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Transactional buffer dump:    [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Streaming metrics dump: OracleStreamingChangeEventSourceMetrics{currentScn=null, oldestScn=null, committedScn=null, offsetScn=null, logMinerQueryCount=0, totalProcessedRows=0, totalCapturedDmlCount=0, totalDurationOfFetchingQuery=PT0S, lastCapturedDmlCount=0, lastDurationOfFetchingQuery=PT0S, maxCapturedDmlCount=0, maxDurationOfFetchingQuery=PT0S, totalBatchProcessingDuration=PT0S, lastBatchProcessingDuration=PT0S, maxBatchProcessingDuration=PT0S, maxBatchProcessingThroughput=0, currentLogFileName=null, minLogFilesMined=0, maxLogFilesMined=0, redoLogStatus=null, switchCounter=0, batchSize=20000, millisecondToSleepBetweenMiningQuery=1000, recordMiningHistory=false, hoursToKeepTransaction=0, networkConnectionProblemsCounter0, batchSizeDefault=20000, batchSizeMin=1000, batchSizeMax=100000, sleepTimeDefault=1000, sleepTimeMin=0, sleepTimeMax=3000, sleepTimeIncrement=200, totalParseTime=PT0S, totalStartLogMiningSessionDuration=PT0S, lastStartLogMiningSessionDuration=PT0S, maxStartLogMiningSessionDuration=PT0S, totalProcessTime=PT0S, minBatchProcessTime=PT0S, maxBatchProcessTime=PT0S, totalResultSetNextTime=PT0S, lagFromTheSource=DurationPT0S, maxLagFromTheSourceDuration=PT0S, minLagFromTheSourceDuration=PT0S, lastCommitDuration=PT0S, maxCommitDuration=PT0S, activeTransactions=0, rolledBackTransactions=0, committedTransactions=0, abandonedTransactionIds=[], rolledbackTransactionIds=[], registeredDmlCount=0, committedDmlCount=0, errorCount=1, warningCount=0, scnFreezeCount=0}   [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Finished streaming   [io.debezium.pipeline.ChangeEventSourceCoordinator]

解决:

  /**
   * oracle 开启补充日志
   */
  // 查询补充日志和强制日志
  // select supplemental_log_data_all,force_logging from v$database;
  
  //alter database add supplemental log data(all) columns;
  //alter database force logging;
  
  /*
   * SQL> select supplemental_log_data_all,force_logging from v$database;

	SUP FORCE_LOGGING
	--- ---------------------------------------
	NO  NO
	
	SQL> alter database add supplemental log data(all) columns;
	
	Database altered.
	
	SQL> alter database force logging;
	
	Database altered.
	
	SQL> select supplemental_log_data_all,force_logging from v$database;
	
	SUP FORCE_LOGGING
	--- ---------------------------------------
	YES YES

   */
相关推荐
Yz98764 分钟前
Hadoop里面MapReduce的序列化与Java序列化比较
java·大数据·jvm·hadoop·分布式·mapreduce·big data
Fan1 小时前
Elasticsearch 下载安装及使用总结
大数据·elasticsearch·jenkins
爬山算法2 小时前
Oracle(129) 如何使用闪回归档(Flashback Archive)?
数据库·oracle
m0_713344852 小时前
新能源汽车数据大全(产销数据\充电桩\专利等)
大数据·人工智能·新能源汽车
goTsHgo3 小时前
从底层原理上解释 ClickHouse 的索引
大数据·clickhouse
Yz98763 小时前
Hadoop-MapReduce的 原理 | 块和片 | Shuffle 过程 | Combiner
大数据·数据库·数据仓库·hadoop·mapreduce·big data
新榜有数3 小时前
品牌建设是什么?怎么做好品牌建设?
大数据·矩阵·数据分析·新媒体运营·流量运营·媒体·内容运营
porkczr3 小时前
oracle rac多个实例就相当于多个数据库系统程序
数据库·oracle
好记性+烂笔头3 小时前
Flink提交任务
大数据·flink
Apache Flink3 小时前
Apache Flink 流批融合技术介绍
人工智能·flink·apache