[Webinar] Mastering Kafka Security Across Hybrid Environments → Register Now

Modernize Hybrid and Multi-Cloud Environments with Treehouse Software and Confluent

Written By

At Treehouse Software, when we speak with customers who are planning to modernize their enterprise mainframe systems, there’s a common theme:  they are faced with decades of mission-critical and historical legacy mainframe data in disparate databases, as well as a variety of other data stores inherited through mergers, acquisitions, and other company growth scenarios. Many applications, connections, databases, and stores are located in different on-premises systems, or on multiple cloud platforms. Customers say they often find themselves in the middle of an organically formed, complex, and multi-cloud environment with little historical context, and they are trying to connect and integrate these systems as best they can. Generally, those integrations tend to be point-to-point, brittle, and tend not to scale with growth.

We’ve seen a growing interest in setting up mainframe-to-Confluent data pipelines without necessarily having a final target fully thought out.  At first glance, this can come across as strange – rather like building a bridge to nowhere, as Confluent is often not considered a final datastore, but simply an event streaming platform. But in the bigger, longer-term picture, an enterprise can keep its options open by propagating data to a highly reliable, very scalable platform like Confluent that can be “subscribed to” by any number of current or yet-to-be-invented ETL toolsets and target datastores. Many firms have found success with this approach of using their event brokers as a central nervous system for business-critical data.

With that said, a site that is doing mainframe-to-Confluent propagation does ultimately need to be able to pull the data from Confluent and land it into a viable target datastore. So Treehouse Software’s current work—enabling the targeting of DynamoDB, Cosmos DB, Snowflake, and others as destinations—is seeing increased popularity among new and existing customers.

The most common Mainframe-to-Confluent use cases

Customers want to modernize applications on cloud and/or open systems without disrupting the existing critical work on legacy systems. They also want to bring together, view, and manage data from applications, databases, data warehouses, etc. that have been spread over many vastly different systems.

The Treehouse and Confluent Solution: Avoid replicating the same complexity to newer systems

Confluent allows customers to eliminate the need for point-to-point interconnections and replace brittle point-to-point interconnections with a real-time, global data plane that connects all of the systems, applications, datastores, and environments that make up an enterprise. That’s possible regardless of whether systems are running on-prem, in the cloud, or a combination of both.

Greg DeMichillie, Vice President of Product and Solutions Marketing at Confluent discusses transitioning to a hybrid or multicloud architecture:

For those customers looking to move mainframe data to Confluent, Treehouse Software’s tcVISION is the mainframe data connector that performs real-time synchronization of data sources to Confluent Platform or Cloud, allowing for rapid data movement to newer data sinks/target platforms on AWS, Azure, Google Cloud, and other services.

Additionally, tcVISION supports many mainframe data sources for both online and offline scenarios. Data can be replicated from IBM DB2 z/OS, DB2 z/VSE, VSAM, IMS/DB, CA IDMS, CA DATACOM, or Software AG ADABAS. tcVISION can replicate data to many targets including Confluent Platform or Cloud. To learn more, see the complete list of supported tcVISION sources and targets. Here’s a look at the architecture that’s created with Confluent and tcVISION:

Learn more in this post on enterprise change data capture (CDC) to Kafka with tcVISION and Confluent.

With tcVISION’s groundbreaking mainframe CDC connector and Confluent’s ability to serve as the multi-tenant data hub, it’s possible to aggregate data from multiple sources and have data published into various Kafka topics. End-to-end data in motion under a simplified hybrid and multicloud architecture enables enterprise customers to oversee data pipelines as well as manage policy and governance.

  • Ram works as a Staff Solutions Engineer at Confluent. He has a wide array of experience in NoSQL databases, Filesystems, Distributed Systems and Apache Kafka. His current interests are in helping customers adopt realtime event streaming technologies using Kafka. He supports various industry verticals ranging from large retailers, healthcare, telecom, and utilities companies towards their digital modernization journey.

  • Joseph Brady is the director of business development and cloud alliance leader at Treehouse Software.  He has been with Treehouse since 1996, and leads enterprise mainframe modernization strategies and partnerships with some of the largest Cloud technology companies and systems integrators in the world.

Did you like this blog post? Share it now

MiFID II: Data Streaming for Post-Trade Reporting

The Markets in Financial Instruments Directive II (MiFID II) came into effect in January 2018, aiming to improve the competitiveness and transparency of European financial markets. As part of this, financial institutions are obligated to report details of trades and transactions (both equity and...