flink-cdc实战之oracle问题记录01

记录问题,温暖你我,上台

欢迎点赞留言关注

复制代码
2024-01-26 11:02:56,168 ERROR  Oracle|oracle_logminer|streaming  Mining session stopped due to the {}   [io.debezium.connector.oracle.logminer.LogMinerHelper]
io.debezium.DebeziumException: Supplemental logging not configured for table ORCL.AA.A.  Use command: ALTER TABLE AA.A ADD SUPPLEMENTAL LOG DATA (ALL) COLUMNS
	at io.debezium.connector.oracle.logminer.LogMinerHelper.checkSupplementalLogging(LogMinerHelper.java:407)
	at io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource.execute(LogMinerStreamingChangeEventSource.java:132)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:152)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:119)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:834)
2024-01-26 11:02:56,170 ERROR  Oracle|oracle_logminer|streaming  Producer failure   [io.debezium.pipeline.ErrorHandler]
io.debezium.DebeziumException: Supplemental logging not configured for table ORCL.AA.A.  Use command: ALTER TABLE AA.A ADD SUPPLEMENTAL LOG DATA (ALL) COLUMNS
	at io.debezium.connector.oracle.logminer.LogMinerHelper.checkSupplementalLogging(LogMinerHelper.java:407)
	at io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource.execute(LogMinerStreamingChangeEventSource.java:132)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:152)
	at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:119)
	at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
	at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at java.base/java.lang.Thread.run(Thread.java:834)
2024-01-26 11:02:56,170 INFO   Oracle|oracle_logminer|streaming  startScn=3460974, endScn=null, offsetContext.getScn()=3460974   [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Transactional buffer dump:    [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Streaming metrics dump: OracleStreamingChangeEventSourceMetrics{currentScn=null, oldestScn=null, committedScn=null, offsetScn=null, logMinerQueryCount=0, totalProcessedRows=0, totalCapturedDmlCount=0, totalDurationOfFetchingQuery=PT0S, lastCapturedDmlCount=0, lastDurationOfFetchingQuery=PT0S, maxCapturedDmlCount=0, maxDurationOfFetchingQuery=PT0S, totalBatchProcessingDuration=PT0S, lastBatchProcessingDuration=PT0S, maxBatchProcessingDuration=PT0S, maxBatchProcessingThroughput=0, currentLogFileName=null, minLogFilesMined=0, maxLogFilesMined=0, redoLogStatus=null, switchCounter=0, batchSize=20000, millisecondToSleepBetweenMiningQuery=1000, recordMiningHistory=false, hoursToKeepTransaction=0, networkConnectionProblemsCounter0, batchSizeDefault=20000, batchSizeMin=1000, batchSizeMax=100000, sleepTimeDefault=1000, sleepTimeMin=0, sleepTimeMax=3000, sleepTimeIncrement=200, totalParseTime=PT0S, totalStartLogMiningSessionDuration=PT0S, lastStartLogMiningSessionDuration=PT0S, maxStartLogMiningSessionDuration=PT0S, totalProcessTime=PT0S, minBatchProcessTime=PT0S, maxBatchProcessTime=PT0S, totalResultSetNextTime=PT0S, lagFromTheSource=DurationPT0S, maxLagFromTheSourceDuration=PT0S, minLagFromTheSourceDuration=PT0S, lastCommitDuration=PT0S, maxCommitDuration=PT0S, activeTransactions=0, rolledBackTransactions=0, committedTransactions=0, abandonedTransactionIds=[], rolledbackTransactionIds=[], registeredDmlCount=0, committedDmlCount=0, errorCount=1, warningCount=0, scnFreezeCount=0}   [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]
2024-01-26 11:02:56,171 INFO   Oracle|oracle_logminer|streaming  Finished streaming   [io.debezium.pipeline.ChangeEventSourceCoordinator]

解决:

复制代码
  /**
   * oracle 开启补充日志
   */
  // 查询补充日志和强制日志
  // select supplemental_log_data_all,force_logging from v$database;
  
  //alter database add supplemental log data(all) columns;
  //alter database force logging;
  
  /*
   * SQL> select supplemental_log_data_all,force_logging from v$database;

	SUP FORCE_LOGGING
	--- ---------------------------------------
	NO  NO
	
	SQL> alter database add supplemental log data(all) columns;
	
	Database altered.
	
	SQL> alter database force logging;
	
	Database altered.
	
	SQL> select supplemental_log_data_all,force_logging from v$database;
	
	SUP FORCE_LOGGING
	--- ---------------------------------------
	YES YES

   */
相关推荐
人道领域1 小时前
AI抢人大战:谁在收割你的红包
大数据·人工智能·算法
Goat恶霸詹姆斯1 小时前
mysql常用语句
数据库·mysql·oracle
qq_12498707531 小时前
基于Hadoop的信贷风险评估的数据可视化分析与预测系统的设计与实现(源码+论文+部署+安装)
大数据·人工智能·hadoop·分布式·信息可视化·毕业设计·计算机毕业设计
Hello.Reader1 小时前
Flink 使用 Amazon S3 读写、Checkpoint、插件选择与性能优化
大数据·flink
xiaowu0802 小时前
C# 拆解 “显式接口实现 + 子类强类型扩展” 的设计思想
数据库·oracle
零售ERP菜鸟2 小时前
范式革命:从“信息化”到“数字化”的本质跃迁
大数据·人工智能·职场和发展·创业创新·学习方法·业界资讯
Hello.Reader2 小时前
Flink 对接 Google Cloud Storage(GCS)读写、Checkpoint、插件安装与生产配置指南
大数据·flink
Hello.Reader2 小时前
Flink Kubernetes HA(高可用)实战原理、前置条件、配置项与数据保留机制
贪心算法·flink·kubernetes
Apple_羊先森3 小时前
ORACLE数据库巡检SQL脚本--19、磁盘读次数最高的前5条SQL语句
数据库·sql·oracle
浪子小院3 小时前
ModelEngine 智能体全流程开发实战:从 0 到 1 搭建多协作办公助手
大数据·人工智能