Hi team, I am getting the below error after whitelisting a new table in mysql connector (version 1.1Final)
org.apache.kafka.connect.errors.ConnectException: Encountered change event for table dominus.payment whose schema isn't known to this connector at io.debezium.connector.mysql.AbstractReader.wrap(AbstractReader.java:230) at io.debezium.connector.mysql.AbstractReader.failed(AbstractReader.java:207) at io.debezium.connector.mysql.BinlogReader.handleEvent(BinlogReader.java:536) at com.github.shyiko.mysql.binlog.BinaryLogClient.notifyEventListeners(BinaryLogClient.java:1095) at com.github.shyiko.mysql.binlog.BinaryLogClient.listenForEventPackets(BinaryLogClient.java:943) at com.github.shyiko.mysql.binlog.BinaryLogClient.connect(BinaryLogClient.java:580) at com.github.shyiko.mysql.binlog.BinaryLogClient$7.run(BinaryLogClient.java:825) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.kafka.connect.errors.ConnectException: Encountered change event for table dominus.payment whose schema isn't known to this connector at io.debezium.connector.mysql.BinlogReader.informAboutUnknownTableIfRequired(BinlogReader.java:794) at io.debezium.connector.mysql.BinlogReader.handleUpdateTableMetadata(BinlogReader.java:768) at io.debezium.connector.mysql.BinlogReader.handleEvent(BinlogReader.java:519) ... 5 more
@MrTrustworthy you also faced similar issue earlier, what is the resolution here?
ColumnImplshould neither face up in docs of the actual option nor in actual user docs; it's an implementation detail and no one will know what this is about
Hello, the 1.7 and 1.8 version gets :
Error: Could not find or load main class org.apache.kafka.connect.cli.ConnectDistributed
I run with the default config
docker run -it --name connect -p 8083:8083 -e GROUP_ID=1 -e CONFIG_STORAGE_TOPIC=my-connect-configs -e OFFSET_STORAGE_TOPIC=my-connect-offsets -e ADVERTISED_HOST_NAME=22.214.171.124 debezium/connect
what is the wrong?
Hi Jiri, I wanted to use kafka connect in different machines( with different IP's) , I started zookeeper, server and kafka connect with debezium and realised the connectors I connect in 1st machine is syncing with the other machine, I can't figure out why ( I am using same DB from SQL Server in org), is it because of CDC of same DB or there could be some other reason?
I am using
advertised.host.name as localhost should I change it too... I am using distributed mode configuration... Is there any article that I can follow to use different machines in Debezium and sync data accordingly.. ( I want to use more than 1 machines for fault tolerence and scaling.)
Any help will be really appreciated, thanks in advance :)