zjcnb opened a new issue, #18075: URL: https://github.com/apache/shardingsphere/issues/18075
## Bug Report **For English only**, other languages will not accept. Before report a bug, make sure you have: - Searched open and closed [GitHub issues](https://github.com/apache/shardingsphere/issues). - Read documentation: [ShardingSphere Doc](https://shardingsphere.apache.org/document/current/en/overview). Please pay attention on issues you submitted, because we maybe need more details. If no response anymore and we cannot reproduce it on current information, we will **close it**. Please answer these questions before submitting your issue. Thanks! ### Which version of ShardingSphere did you use? `master` ### Which project did you use? ShardingSphere-JDBC or ShardingSphere-Proxy? `ShardingSphere-Proxy` ### Expected behavior No error msg ### Actual behavior ``` [INFO ] 2022-05-30 20:50:14.838 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:14.838 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=2, finishedByBreak=true [INFO ] 2022-05-30 20:50:14.838 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-4] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_0.t_order_0#0 [INFO ] 2022-05-30 20:50:14.839 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-5] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_0.t_order_item_0#0 [INFO ] 2022-05-30 20:50:14.839 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:14.839 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:14.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:14.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:14.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=3, finishedByBreak=true [INFO ] 2022-05-30 20:50:14.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=3, finishedByBreak=true [INFO ] 2022-05-30 20:50:14.842 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-6] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_0.t_order_item_2#0 [INFO ] 2022-05-30 20:50:14.842 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-7] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_0.t_order_2#0 [INFO ] 2022-05-30 20:50:14.842 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:14.842 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:14.843 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-6] o.a.s.d.p.s.r.RuleAlteredJobScheduler - onSuccess, all inventory tasks finished. [INFO ] 2022-05-30 20:50:14.843 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-7] o.a.s.d.p.s.r.RuleAlteredJobScheduler - onSuccess, all inventory tasks finished. [INFO ] 2022-05-30 20:50:14.843 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-6] o.a.s.d.p.s.r.RuleAlteredJobScheduler - -------------- Start incremental task -------------- [INFO ] 2022-05-30 20:50:14.844 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-7] o.a.s.d.p.s.r.RuleAlteredJobScheduler - job status already EXECUTE_INCREMENTAL_TASK, ignore [INFO ] 2022-05-30 20:50:14.844 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.IncrementalTask@529fa392 [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-1] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@647941e6 [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-2] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@28e949f3 [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-3] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@74293cbf [INFO ] 2022-05-30 20:50:14.845 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:14.846 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLIncrementalDumper@7904bb3e [INFO ] 2022-05-30 20:50:14.846 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.m.i.MySQLIncrementalDumper - incremental dump, jdbcUrl=jdbc:mysql://192.168.10.18:33307/original_1?rewriteBatchedStatements=true&serverTimezone=UTC&yearIsDateType=false&useSSL=false [ERROR] 2022-05-30 20:50:14.906 [nioEventLoopGroup-4-1] o.a.s.d.p.m.i.client.MySQLClient - protocol resolution error io.netty.handler.codec.DecoderException: java.lang.UnsupportedOperationException: Do not parse binlog event fully, eventHeader: MySQLBinlogEventHeader(timestamp=1653560367, eventType=15, serverId=133307, eventSize=121, logPos=0, flags=0), remaining packet 3a9d9af2 at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:480) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:279) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:327) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:314) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:435) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:279) at io.netty.handler.codec.ByteToMessageCodec.channelRead(ByteToMessageCodec.java:103) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:722) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:658) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:584) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:986) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.lang.UnsupportedOperationException: Do not parse binlog event fully, eventHeader: MySQLBinlogEventHeader(timestamp=1653560367, eventType=15, serverId=133307, eventSize=121, logPos=0, flags=0), remaining packet 3a9d9af2 at org.apache.shardingsphere.data.pipeline.mysql.ingest.client.netty.MySQLBinlogEventPacketDecoder.decode(MySQLBinlogEventPacketDecoder.java:89) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:510) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:449) ... 25 common frames omitted [INFO ] 2022-05-30 20:50:14.906 [nioEventLoopGroup-4-1] o.a.s.d.p.m.i.client.MySQLClient - reconnect mysql client. [INFO ] 2022-05-30 20:50:16.496 [Thread-9] o.a.s.d.p.s.r.RuleAlteredJobPreparer - unlocked, lockName=prepare-0130317c30317c3054317c74657374 [INFO ] 2022-05-30 20:50:16.517 [Thread-9] o.a.s.d.p.c.m.l.PipelineTableMetaDataLoader - loadTableMetaData, schemaNameFinal=null, tableNamePattern=t_order_1, result={t_order_1=PipelineTableMetaData(name=t_order_1, columnMetaDataMap={order_id=PipelineColumnMetaData(ordinalPosition=1, name=order_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=true), user_id=PipelineColumnMetaData(ordinalPosition=2, name=user_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), status=PipelineColumnMetaData(ordinalPosition=3, name=status, dataType=12, dataTypeName=VARCHAR, nullable=true, primaryKey=false)}, columnNames=[order_id, user_id, status], primaryKeyColumns=[order_id], uniqueIndexes=[org.apache.shardingsphere.data.pipeline.core.metadata.model.PipelineIndexMetaData@106fc0bf])}, cost time=14 ms [INFO ] 2022-05-30 20:50:16.519 [Thread-9] o.a.s.d.p.s.r.p.InventoryTaskSplitter - getPositionByPrimaryKeyRange, endId is 0, break, tableName=t_order_1, primaryKey=order_id, beginId=6 [INFO ] 2022-05-30 20:50:16.530 [Thread-9] o.a.s.d.p.c.m.l.PipelineTableMetaDataLoader - loadTableMetaData, schemaNameFinal=null, tableNamePattern=t_order_3, result={t_order_3=PipelineTableMetaData(name=t_order_3, columnMetaDataMap={order_id=PipelineColumnMetaData(ordinalPosition=1, name=order_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=true), user_id=PipelineColumnMetaData(ordinalPosition=2, name=user_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), status=PipelineColumnMetaData(ordinalPosition=3, name=status, dataType=12, dataTypeName=VARCHAR, nullable=true, primaryKey=false)}, columnNames=[order_id, user_id, status], primaryKeyColumns=[order_id], uniqueIndexes=[org.apache.shardingsphere.data.pipeline.core.metadata.model.PipelineIndexMetaData@4085a818])}, cost time=11 ms [INFO ] 2022-05-30 20:50:16.531 [Thread-9] o.a.s.d.p.s.r.p.InventoryTaskSplitter - getPositionByPrimaryKeyRange, endId is 0, break, tableName=t_order_3, primaryKey=order_id, beginId=4 [INFO ] 2022-05-30 20:50:16.545 [Thread-9] o.a.s.d.p.c.m.l.PipelineTableMetaDataLoader - loadTableMetaData, schemaNameFinal=null, tableNamePattern=t_order_item_1, result={t_order_item_1=PipelineTableMetaData(name=t_order_item_1, columnMetaDataMap={item_id=PipelineColumnMetaData(ordinalPosition=1, name=item_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=true), order_id=PipelineColumnMetaData(ordinalPosition=2, name=order_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), user_id=PipelineColumnMetaData(ordinalPosition=3, name=user_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), status=PipelineColumnMetaData(ordinalPosition=4, name=status, dataType=12, dataTypeName=VARCHAR, nullable=true, primaryKey=false), creation_date=PipelineColumnMetaData(ordinalPosition=5, name=creation_date, dataType=91, dataTypeName=DATE, nullable=true, primaryKey=false)}, columnNames=[item_id, order_id, user_id, status, creation_date], primaryKeyColumns=[i tem_id], uniqueIndexes=[org.apache.shardingsphere.data.pipeline.core.metadata.model.PipelineIndexMetaData@25aff154])}, cost time=14 ms [INFO ] 2022-05-30 20:50:16.546 [Thread-9] o.a.s.d.p.s.r.p.InventoryTaskSplitter - getPositionByPrimaryKeyRange, endId is 0, break, tableName=t_order_item_1, primaryKey=item_id, beginId=6 [INFO ] 2022-05-30 20:50:16.559 [Thread-9] o.a.s.d.p.c.m.l.PipelineTableMetaDataLoader - loadTableMetaData, schemaNameFinal=null, tableNamePattern=t_order_item_3, result={t_order_item_3=PipelineTableMetaData(name=t_order_item_3, columnMetaDataMap={item_id=PipelineColumnMetaData(ordinalPosition=1, name=item_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=true), order_id=PipelineColumnMetaData(ordinalPosition=2, name=order_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), user_id=PipelineColumnMetaData(ordinalPosition=3, name=user_id, dataType=4, dataTypeName=INT, nullable=false, primaryKey=false), status=PipelineColumnMetaData(ordinalPosition=4, name=status, dataType=12, dataTypeName=VARCHAR, nullable=true, primaryKey=false), creation_date=PipelineColumnMetaData(ordinalPosition=5, name=creation_date, dataType=91, dataTypeName=DATE, nullable=true, primaryKey=false)}, columnNames=[item_id, order_id, user_id, status, creation_date], primaryKeyColumns=[i tem_id], uniqueIndexes=[org.apache.shardingsphere.data.pipeline.core.metadata.model.PipelineIndexMetaData@23424faa])}, cost time=13 ms [INFO ] 2022-05-30 20:50:16.560 [Thread-9] o.a.s.d.p.s.r.p.InventoryTaskSplitter - getPositionByPrimaryKeyRange, endId is 0, break, tableName=t_order_item_3, primaryKey=item_id, beginId=4 [INFO ] 2022-05-30 20:50:16.560 [Thread-9] o.a.s.d.p.s.r.RuleAlteredJobPreparer - prepare, jobId=0130317c30317c3054317c74657374, shardingItem=1, inventoryTasks=[InventoryTask(taskId=ds_1.t_order_1#0, position=i,0,5), InventoryTask(taskId=ds_1.t_order_3#0, position=i,0,3), InventoryTask(taskId=ds_1.t_order_item_1#0, position=i,0,5), InventoryTask(taskId=ds_1.t_order_item_3#0, position=i,0,3)], incrementalTasks=[IncrementalTask(taskId=ds_1)] [INFO ] 2022-05-30 20:50:16.563 [Thread-9] o.a.s.d.p.s.r.RuleAlteredJobScheduler - -------------- Start inventory task -------------- [INFO ] 2022-05-30 20:50:16.564 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.InventoryTask@77e3cc33 [INFO ] 2022-05-30 20:50:16.564 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.InventoryTask@7b12f15d [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.InventoryTask@45a33fa9 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLInventoryDumper@7d1ba743 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@6747aa25 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, uniqueKeyDataType=4, firstSQL=SELECT * FROM `t_order_1` WHERE `order_id` >= ? AND `order_id` <= ? ORDER BY `order_id` ASC LIMIT ?, laterSQL=SELECT * FROM `t_order_1` WHERE `order_id` > ? AND `order_id` <= ? ORDER BY `order_id` ASC LIMIT ?, position=i,0,5 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.InventoryTask@5b48906b [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLInventoryDumper@71ea9a2d [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, uniqueKeyDataType=4, firstSQL=SELECT * FROM `t_order_3` WHERE `order_id` >= ? AND `order_id` <= ? ORDER BY `order_id` ASC LIMIT ?, laterSQL=SELECT * FROM `t_order_3` WHERE `order_id` > ? AND `order_id` <= ? ORDER BY `order_id` ASC LIMIT ?, position=i,0,3 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLInventoryDumper@4f5b54d9 [INFO ] 2022-05-30 20:50:16.565 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@7d08e610 [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, uniqueKeyDataType=4, firstSQL=SELECT * FROM `t_order_item_1` WHERE `item_id` >= ? AND `item_id` <= ? ORDER BY `item_id` ASC LIMIT ?, laterSQL=SELECT * FROM `t_order_item_1` WHERE `item_id` > ? AND `item_id` <= ? ORDER BY `item_id` ASC LIMIT ?, position=i,0,5 [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLInventoryDumper@51b0fb1b [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, uniqueKeyDataType=4, firstSQL=SELECT * FROM `t_order_item_3` WHERE `item_id` >= ? AND `item_id` <= ? ORDER BY `item_id` ASC LIMIT ?, laterSQL=SELECT * FROM `t_order_item_3` WHERE `item_id` > ? AND `item_id` <= ? ORDER BY `item_id` ASC LIMIT ?, position=i,0,3 [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@388d4882 [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@21fb12ed [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:16.566 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:16.567 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump done, round=3, maxUniqueKeyValue=Optional.empty [INFO ] 2022-05-30 20:50:16.567 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, before put FinishedRecord [INFO ] 2022-05-30 20:50:16.570 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump done, round=3, maxUniqueKeyValue=Optional.empty [INFO ] 2022-05-30 20:50:16.570 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, before put FinishedRecord [INFO ] 2022-05-30 20:50:16.572 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump done, round=3, maxUniqueKeyValue=Optional.empty [INFO ] 2022-05-30 20:50:16.572 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, before put FinishedRecord [INFO ] 2022-05-30 20:50:16.572 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump done, round=3, maxUniqueKeyValue=Optional.empty [INFO ] 2022-05-30 20:50:16.572 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.d.AbstractInventoryDumper - inventory dump, before put FinishedRecord [INFO ] 2022-05-30 20:50:29.837 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:29.837 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=2, finishedByBreak=true [INFO ] 2022-05-30 20:50:29.838 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-4] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_1.t_order_3#0 [INFO ] 2022-05-30 20:50:29.838 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-1] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:29.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:29.841 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=2, finishedByBreak=true [INFO ] 2022-05-30 20:50:29.842 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-5] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_1.t_order_item_3#0 [INFO ] 2022-05-30 20:50:29.842 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-3] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:29.844 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:29.844 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.AbstractImporter - write, get FinishedRecord, break [INFO ] 2022-05-30 20:50:29.844 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=3, finishedByBreak=true [INFO ] 2022-05-30 20:50:29.844 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-0] o.a.s.d.p.c.i.AbstractImporter - importer write done, rowCount=3, finishedByBreak=true [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-6] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_1.t_order_item_1#0 [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-2] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Importer-0130317c30317c3054317c74657374-7] o.a.s.d.p.core.task.InventoryTask - importer onSuccess, taskId=ds_1.t_order_1#0 [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-0] o.a.s.d.p.core.task.InventoryTask - importer future done [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-6] o.a.s.d.p.s.r.RuleAlteredJobScheduler - onSuccess, all inventory tasks finished. [INFO ] 2022-05-30 20:50:29.845 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-6] o.a.s.d.p.s.r.RuleAlteredJobScheduler - -------------- Start incremental task -------------- [INFO ] 2022-05-30 20:50:29.846 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-7] o.a.s.d.p.s.r.RuleAlteredJobScheduler - onSuccess, all inventory tasks finished. [INFO ] 2022-05-30 20:50:29.846 [ShardingSphere-Scaling-Inventory-0130317c30317c3054317c74657374-7] o.a.s.d.p.s.r.RuleAlteredJobScheduler - job status already EXECUTE_INCREMENTAL_TASK, ignore [INFO ] 2022-05-30 20:50:29.846 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.core.task.IncrementalTask@afed38c [INFO ] 2022-05-30 20:50:29.846 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-1] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@6f6ed929 [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-1] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-2] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@58a5b9e2 [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-2] o.a.s.d.p.c.i.AbstractImporter - importer write [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.ingest.MySQLIncrementalDumper@391aef07 [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-0] o.a.s.d.p.m.i.MySQLIncrementalDumper - incremental dump, jdbcUrl=jdbc:mysql://192.168.10.18:33307/original_2?rewriteBatchedStatements=true&serverTimezone=UTC&yearIsDateType=false&useSSL=false [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-3] o.a.s.d.p.a.e.AbstractLifecycleExecutor - start lifecycle executor: org.apache.shardingsphere.data.pipeline.mysql.importer.MySQLImporter@4d753e7c [INFO ] 2022-05-30 20:50:29.847 [ShardingSphere-Scaling-Incremental-0130317c30317c3054317c74657374-3] o.a.s.d.p.c.i.AbstractImporter - importer write [ERROR] 2022-05-30 20:50:29.863 [nioEventLoopGroup-5-1] o.a.s.d.p.m.i.client.MySQLClient - protocol resolution error io.netty.handler.codec.DecoderException: java.lang.UnsupportedOperationException: Do not parse binlog event fully, eventHeader: MySQLBinlogEventHeader(timestamp=1653560367, eventType=15, serverId=133307, eventSize=121, logPos=0, flags=0), remaining packet 3a9d9af2 at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:480) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:279) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:327) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:314) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:435) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:279) at io.netty.handler.codec.ByteToMessageCodec.channelRead(ByteToMessageCodec.java:103) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:722) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:658) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:584) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:986) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.lang.UnsupportedOperationException: Do not parse binlog event fully, eventHeader: MySQLBinlogEventHeader(timestamp=1653560367, eventType=15, serverId=133307, eventSize=121, logPos=0, flags=0), remaining packet 3a9d9af2 at org.apache.shardingsphere.data.pipeline.mysql.ingest.client.netty.MySQLBinlogEventPacketDecoder.decode(MySQLBinlogEventPacketDecoder.java:89) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:510) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:449) ... 25 common frames omitted [INFO ] 2022-05-30 20:50:29.863 [nioEventLoopGroup-5-1] o.a.s.d.p.m.i.client.MySQLClient - reconnect mysql client. [INFO ] 2022-05-30 20:50:45.042 [_finished_check_Worker-1] o.a.s.d.p.core.job.FinishedCheckJob - completionDetector not configured, auto switch will not be enabled. You could query job progress and switch config manually with DistSQL. [INFO ] 2022-05-30 20:50:50.033 [_finished_check_Worker-1] o.a.s.d.p.core.job.FinishedCheckJob - completionDetector not configured, auto switch will not be enabled. You could query job progress and switch config manually with DistSQL. [INFO ] 2022-05-30 20:51:00.222 [_finished_check_Worker-1] o.a.s.d.p.core.job.FinishedCheckJob - completionDetector not configured, auto switch will not be enabled. You could query job progress and switch config manually with DistSQL. [INFO ] 2022-05-30 20:51:10.215 [_finished_check_Worker-1] o.a.s.d.p.core.job.FinishedCheckJob - completionDetector not configured, auto switch will not be enabled. You could query job progress and switch config manually with DistSQL. [INFO ] 2022-05-30 20:51:20.220 [_finished_check_Worker-1] o.a.s.d.p.core.job.FinishedCheckJob - completionDetector not configured, auto switch will not be enabled. You could query job progress and switch config manually with DistSQL. ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
