Our great sponsors
Kafka Connect connectors for Azure Cosmos DB
The Azure Cosmos DB connector allows you to move data between Azure Cosmos DB and Kafka. It’s available as a source as well as a sink. The Azure Cosmos DB Sink connector writes data from a Kafka topic to an Azure Cosmos DB container and the Source connector writes changes from an Azure Cosmos DB container to a Kafka topic. At the time of writing, the connector is in pre-production mode. You can read more about it on the GitHub repo or install/download it from the Confluent Hub.
Getting started with Kafka Connector for Azure Cosmos DB using Docker
git clone [https://github.com/Azure-Samples/cosmosdb-kafka-connect-docker](https://github.com/Azure-Samples/cosmosdb-kafka-connect-docker) cd cosmosdb-kafka-connect-docker
Access the most powerful time series database as a service. Ingest, store, & analyze all types of time series data in a fully-managed, purpose-built database. Keep data forever with low-cost storage and superior data compression.
Connector that generates data for demos
For the remaining scenarios, we will use a producer component to generate records. The Kafka Connect Datagen connector is our friend. It is meant for generating mock data for testing, so let’s put it to good use!
Define and run multi-container applications with Docker
And, Docker Compose which is a tool for defining and running multi-container Docker applications. It will orchestrate all the components required by our setup including Azure Cosmos DB emulator, Kafka, Zookeeper, Kafka connectors etc.
Source repo for Docker's Documentation (by docker)
Having a local development environment is quite handy when trying out a new service or technology. Docker has emerged as the de-facto choice in such cases. It is specially useful in scenarios where you’re trying to integrate multiple services and gives you the ability to to start fresh before each run.
Apache Avro is a data serialization system.
So far we dealt with JSON, a commonly used data format. But, Avro is heavily used in production due to its compact format which leads to better performance and cost savings. To make it easier to deal with Avro data schema, there is Confluent Schema Registry which provides a serving layer for your metadata along with a RESTful interface for storing and retrieving your Avro (as well as JSON and Protobuf schemas). We will use the Docker version for the purposes of this blog post.
Dockerize an Django Based app along Postgres + Redis + ElasticSearch 🚀
3 projects | dev.to | 9 Nov 2022
Ask HN: Go vs. Python (Docker SDK)
3 projects | news.ycombinator.com | 24 Aug 2022
Output colored logs from django through supervisor and docker-compose
2 projects | reddit.com/r/codehunter | 25 Apr 2022
Best way to scripting docker commands?
2 projects | reddit.com/r/docker | 23 Mar 2022
Docker & Docker-compose installation with ansible playbook doesn't work
2 projects | reddit.com/r/ansible | 15 Jan 2022