Skip to content

Mqtt, Kafka, Kafka Connect and MongoDB based datapipeline

Notifications You must be signed in to change notification settings

suarabhsuman27/datapipeline

Repository files navigation

Mqtt, Kafka, Kafka Connect and MongoDB based datapipeline

  • Kafka-connect-utilities directory contains all the necessary libary and configuration for Sink the incoming data to MongoDB. Refer the README file for configuration set up

  • Simulator directory contains two Python based Mqtt Client which keep sending data to the broker at every 5 seconds intervals. These two client will publish data on their respective topic

  • docker-compose.yml contains mongodb and mongo client

  • In this datapipeline application, Apache Kafka has been used. We need to start the Kafka Broker and Kafka-connect. Before starting the kafka connect, set up the configurations mentioned in Kafak-connect-utilities

  • Services directory contains the Springboot based Java application - datapipeline, consumes data published by different client on their topic, perform the data transformations. It transform the data into common data model and then published to the Kafka Broker on a topic. Kafka-connect Mongodb Sink Connect will consume data from that topic and push to MongoDB.

System Architecture

image

About

Mqtt, Kafka, Kafka Connect and MongoDB based datapipeline

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published