Map and persist events from Kafka topics directly to MongoDB collections with ease. Ingest events from your Kakfa topics directly into MongoDB collections, exposing the data to your services for efficient querying, enrichment, and analytics.
Publish data changes from MongoDB into Kafka topics for streaming to consuming apps. Data is captured via Change Streams within the MongoDB cluster and published into Kafka topics. This enables consuming apps to react to data changes in real time using an event-driven programming style.
ao.com, a leading online electrical retailer, uses Kafka to push all data changes from its source databases to MongoDB Atlas. This creates a single source of truth for all customer data to drive new and enhanced applications and business processes including customer service, fraud detection, and GDPR compliance. Employees with appropriate permissions can access customer data from one easy-to-consume operational data layer.
Josh Software, part of a project in India to house more than 100,000 people in affordable smart homes, pushes data from millions of sensors to Kafka, processes it in Apache Spark, and writes the results to MongoDB, which connects the operational and analytical data sets. By streaming data from millions of sensors in near real-time, the project is creating truly smart homes, and citizens can access data via a mobile app to better manage their homes.
AHL, a subsidiary of The Man Group, which is one of the world’s largest hedge fund investment firms, used MongoDB to create a single platform for all of its financial data. The system receives data for up to 150,000 ticks per second from multiple financial sources and writes it to Kafka. Kafka provides both consolidation and buffering of events before they are stored in MongoDB, where the data can be analyzed.
comparethemarket.com, a leading price comparison provider, uses MongoDB as the default operational database across its microservices architecture. While each microservice uses its own MongoDB database, the company needs to maintain synchronization between services, so every application event is written to a Kafka topic. Relevant events are written to MongoDB to enable real-time personalization and optimize the customer experience.
State, an intelligent opinion network connecting people with similar beliefs, writes survey data to MongoDB and leverages MongoDB Change Streams to push database changes into Kafka topics where they are consumed by its user recommendation engine. This engine suggests potentially interesting users and updates instantly as soon as a user contributes a new opinion.
Download it directly from the Confluent website.
This connector is open source and can be downloaded from our GitHub repo.
Learn about the event-driven architecture and how MongoDB can help get you there.
MongoDB and its Connector for Apache Kafka are core to event-driven application, which helps you run your business in real time.
Move data seamlessly from the MQTT protocol into MongoDB time series collections using Apache Kafka.