-
demo-scene
Scripts and samples to support Confluent Demos, Talks, and Blogs. Not all of the examples in this repository are kept up to date. For automated tutorials and QA'd code, see https://github.com/confluentinc/tutorials/
should I be reading a different material for a first Kafka project and working with a different kind of setup? Now, I can't be unbiased on this one ;) One of the things we're doing with Confluent Developer is to try and create a resource for people to learn Kafka from the ground up, whether they ultimately decide to pursue it on Confluent or not. The fundamentals of Kafka that you'll be learning are going to be as applicable whether you're using Apache Kafka self-managed, or Confluent, or AWS' MSK, or whatever else. Personally I'd this stage I'd use whatever setup you find easiest and least friction to your learning journey. As u/louisvell mentioned, /u/stephanemaarek's courses on Udemy are also very popular, if you wanted a "second opinion" on how to approach learning Kafka.
-
CodeRabbit
CodeRabbit: AI Code Reviews for Developers. Revolutionize your code reviews with AI. CodeRabbit offers PR summaries, code walkthroughs, 1-click suggestions, and AST-based analysis. Boost productivity and code quality across all major languages with each PR.
-
I'm reading O'Reilly's "Mastering Kafka Streams and ksqlDB" to start learning Kafka, it was suggested for me on an ad by Confluent. Unsurprisingly it uses Confluent's software throughout the book. One of the first projects is a simple app that does sentiment analysis on tweets. The book uses kafka-console-producer and a sample .json file for the tweets, but for my app I wanted to read actual tweets. To do that I've been reading about Kafka Connect and looking at this repository, but I'm having a hard time understating how to best deploy this for my local setup. So far I've been using docker-compose.yml files provided by the book, which in turn uses Confluent's docker images for kafka, zookeeper, etc. As for this Twitter Connect repository, it seems the recommended way of setting it up is to use Confluent's platform and its CLI tool to automagically install it, which is fine, but I wanted to learn how things work under the hood (to some extend) and if possible not rely so heavily upon Confluent's software. Is it a good idea to just stick with Confluent and the book, or should I be reading a different material for a first Kafka project and working with a different kind of setup? Perhaps I'm getting ahead of myself trying to use Kafka Connect at this point?
-
Related posts
-
Ingesting XML data into Kafka - Option 2: Kafka Connect plus Single Message Transform
-
Show HN: Settling merge or rebase? with Kafka and JavaScript <3
-
Confused and frustrated about Kafka
-
Any suggestions for writing a long thesis about Kafka's scalability, availability and fault-tolerance?
-
What Are Apache Kafka Consumer Group IDs?