Hi Robin, I am a software engineer at a startup. Last year we build a pipeline to sync our postgres data to elasticsearch and cassandra. It was all custom java code with lot of operational handling. Thank you for this video, I am planning to use connect for those pipelines.
Thanks Robin - from your newest fan and subscriber :) I'm really loving all the information coming from Confluent. Doing a top job. We are getting serious about implementing a solution centralized on Kafka (on limited budget) - guess there is just a lot of different ways and means. Will post on the community bit later - but just wondering - off top of your head if you were combining web logs from multiple websites of a similar nature (db schema is same - although as per your suggestion will look into avro) - would you combine all users into 1 topic (perhaps tagging where they originated) or set-up a topic for each website. Ultimately queries are centralized on username, so origination just fyi. Somewhere I heard/read about creating a topic per user - but this did n't seem right (for 10ks of users)
Hi Mark, from what you describe I would definitely collate these into a single topic, since they sound like the same logical entity. One topic per user sounds…unusual.
Thanks Robin. I have question on Plugin_path. you have given while installing the connector. From where that path came? Can i give any path? Where i can find that path to mention in Dockerfile?
Hi, this path comes from wherever you put the JDBC connector when you installed it. This might help: rmoff.net/2020/06/19/how-to-install-connector-plugins-in-kafka-connect/ If you're still stuck then please go to forum.confluent.io/ and ask for further help there. Thanks.
@@rum81 If you look at past talks from Kafka Summit (www.kafka-summit.org/past-events) you'll find lots of examples of companies using Kafka Connect in production.
Hello Robin, I connected azureSQL with kafkaconnect by giving table name,host name,server name. .But not able to specify the db schema name anywhere, is there any way to specify schema name? Because without specifying schema name it is creating new table in db.
In distributed mode, somtimes connect worker throws error about status.storage.topic cleanup.policy should be set to compact. I'm wondering why it throws that error occasionally!? and...Would setting log.cleanup.policy to compact on Kafka broker fix the issue!?
Yes, they should be set to compact - see docs.confluent.io/kafka-connectors/self-managed/userguide.html#kconnect-internal-topics Also head to confluent.io/community/ask-the-community if you have any more questions :)
Hi Robin, Thanks for amazing videos.We are implementing Kafka in our project and when ever I got stuck your videos are helping a lot to clear of the concepts and issues. I have small conceptual doubt. Does Kafka and Kafka connect supports ENUM datatypes . We are facing error like Type cast the data type when syncing data from source table to sync table .
I'm so glad my videos have helped you out :) I don't know the answer to your ENUM question - please ask at forum.confluent.io/ and someone should be able to help. Thanks.
@@rmoff I am trying to test FileStreamSourceConnector (file-source , a preconfigured connector in apache kafka)........ Connector starts successfully and it also fetches the data into the topic..... but when I run kafka consumer , it does not fetches any record...... i am following this document docs.confluent.io/platform/current/connect/quickstart.html Also, I am unable to find such connector under plugin.path...... then how come connctor starts ?
Did you create the database first? If you're still stuck head to forum.confluent.io/ with full details of what you've run and where you're getting the error.
Hi Robin,
I am a software engineer at a startup. Last year we build a pipeline to sync our postgres data to elasticsearch and cassandra. It was all custom java code with lot of operational handling. Thank you for this video, I am planning to use connect for those pipelines.
Hi Robin,
I never write comments on youtube videos, but i deeply want to thankyou for all your work !
Thanks - glad it was useful!
Your examples are always very well chosen. Thanks.
Thanks - glad you've found it useful :)
Thanks Robin - from your newest fan and subscriber :) I'm really loving all the information coming from Confluent. Doing a top job. We are getting serious about implementing a solution centralized on Kafka (on limited budget) - guess there is just a lot of different ways and means. Will post on the community bit later - but just wondering - off top of your head if you were combining web logs from multiple websites of a similar nature (db schema is same - although as per your suggestion will look into avro) - would you combine all users into 1 topic (perhaps tagging where they originated) or set-up a topic for each website. Ultimately queries are centralized on username, so origination just fyi. Somewhere I heard/read about creating a topic per user - but this did n't seem right (for 10ks of users)
Hi Mark, from what you describe I would definitely collate these into a single topic, since they sound like the same logical entity. One topic per user sounds…unusual.
Hai Robin, I am a new subscriber fan here
Thanks Robin. I have question on Plugin_path. you have given while installing the connector. From where that path came? Can i give any path? Where i can find that path to mention in Dockerfile?
Hi, this path comes from wherever you put the JDBC connector when you installed it. This might help: rmoff.net/2020/06/19/how-to-install-connector-plugins-in-kafka-connect/
If you're still stuck then please go to forum.confluent.io/ and ask for further help there. Thanks.
Hi Robin, thanks for this video. I wonder 'mariadb-jdbc-connect' is available in this project. Thanks :)
Hi, if it has a JDBC driver then it's worth trying with the JDBC Source connector, sure.
Thank you Robin!
my pleasure, glad to help :)
@@rmoff can you share links of use of kafka connect in production by companies. Need these examples to propose connect in my organization
@@rum81 If you look at past talks from Kafka Summit (www.kafka-summit.org/past-events) you'll find lots of examples of companies using Kafka Connect in production.
Can you do a video of "how to integrate MQ with Kafka topic through ibmmq source connector?"
Sure, I'll add it to the list :)
Thanks a lot Robin..i am trying to do it..but it is failing..and am not able to see any logs in control center..:(
Hi Robin, is there a source connector for adobe or can we use a json connector as long as the streaming data is in json format?
The best place to ask is www.confluent.io/en-gb/community/ask-the-community/
Hello Robin, I connected azureSQL with kafkaconnect by giving table name,host name,server name. .But not able to specify the db schema name anywhere, is there any way to specify schema name? Because without specifying schema name it is creating new table in db.
hi, please head over to forum.confluent.io/ and ask there :) thanks.
In distributed mode, somtimes connect worker throws error about status.storage.topic cleanup.policy should be set to compact. I'm wondering why it throws that error occasionally!? and...Would setting log.cleanup.policy to compact on Kafka broker fix the issue!?
Yes, they should be set to compact - see docs.confluent.io/kafka-connectors/self-managed/userguide.html#kconnect-internal-topics
Also head to confluent.io/community/ask-the-community if you have any more questions :)
I hope it isn't too late to thank you Robin
Glad it was useful :)
Hi Robin, I facing issue in creating topic in Kafka for decimal data type is store as byte any way to slove that
Hi Ankit, the best place to ask is confluent.io/community/ask-the-community/
Hi Robin,
Thanks for amazing videos.We are implementing Kafka in our project and when ever I got stuck your videos are helping a lot to clear of the concepts and issues.
I have small conceptual doubt.
Does Kafka and Kafka connect supports ENUM datatypes . We are facing error like Type cast the data type when syncing data from source table to sync table .
I'm so glad my videos have helped you out :)
I don't know the answer to your ENUM question - please ask at forum.confluent.io/ and someone should be able to help. Thanks.
The key format 'AVRO' is not currently supported. - when using FOEMAT='AVRO' in the KSQL
You need to upgrade to a more recent version of ksqlDB.
Can you share any documents for msk as sink connectors
hi, the best place to get help is at www.confluent.io/en-gb/community/ask-the-community/ :)
Hey Robin, thanks for this video. But could u pls guide us first on how to start apache kafka connect? And how to check if it is already running.
You can find good info on running Kafka Connect here: docs.confluent.io/platform/current/connect/userguide.html#connect-userguide-standalone-config
@@rmoff I am trying to test FileStreamSourceConnector (file-source , a preconfigured connector in apache kafka)........ Connector starts successfully and it also fetches the data into the topic..... but when I run kafka consumer , it does not fetches any record...... i am following this document docs.confluent.io/platform/current/connect/quickstart.html
Also, I am unable to find such connector under plugin.path...... then how come connctor starts ?
I'm getting
ERROR 1049 (42000): Unknown database 'demo'
while trying to connect to mysql...
Did you create the database first? If you're still stuck head to forum.confluent.io/ with full details of what you've run and where you're getting the error.
Hi Robin,
How can we include the json schema in the message, when field is an array of objects ? I don't have the option to use avro.
Hi, can you post this at forum.confluent.io/ and hopefully someone will be able to help there :)