Save 25% or More on Your Kafka Costs | Take the Confluent Cost Savings Challenge
Many companies collect and store their data in various data stores and use a number of business applications and services to access, analyze and act on their data. Pulling all the data from disparate sources is difficult to manage, inefficient and ineffective in producing results. Event streaming and stream processing changes this paradigm. By enabling robust and reactive data pipelines between all your data stores, apps and services, you can make real-time decisions that are critical to your business. In this online talk, we introduce Apache Kafka® and the MongoDB connector for Kafka, and demonstrate a real world stock trading use case that joins heterogeneous data sources to find the moving average of securities using Apache Kafka and MongoDB.