RDBMS and Apache Geode Data Movement - Heather Riddle & Paul Warren, HCSC

December 10, 2017

Extract, transform, load (ETL) has always been complex and expensive for moving massive data sets from one data source to another. This is especially true if the source system is a traditional RDBMS with complicated relationships between tables. Most of the time, traditional ETL processes are implemented with batch, monolithic, and tightly coupled approaches. As the result, traditional ETL processes are often considered fragile, hard to maintain, not easy to tune, and often introduce high data latency between source and destination systems. In this session, Paul and Heather will cover how to create cloud-native event driven microservices (ETL pipeline) for RDBMS and Apache Geode by using Cloud Foundry, Spring Cloud Stream, and RabbitMQ/Kafka. The pipelines can handle high volume data sets and complex database queries, yet with low data latency between the source RDBMS and Apache Geode. In addition, the design is highly tunable and scalable. The session will also cover analysis of performance metrics based on the implementations of real world use cases. Slides: TBA Paul Warren, Senior Engineer, HCSC Heather Riddle, Senior Engineer, HCSC Filmed at SpringOne Platform 2017

Previous
How to Build Modern Data Pipelines with Pivotal GemFire and Spring Cloud Data Flow
How to Build Modern Data Pipelines with Pivotal GemFire and Spring Cloud Data Flow

Getting data from point A to B in a timely way is often easier said than done. Learn how to breakdown data ...

Next
An Introduction to Look-Aside Caching
An Introduction to Look-Aside Caching

Learn the basics of look-aside caching, how it works, when to use it and how it differs from inline caching...