Building Data Workflows with Luigi and Kubernetes

Manage complex data pipelines and seamlessly scale them on-demand

Nar Kumar Chhantyal

Architecture Big Data Data Distributed Systems Scaling

See in schedule Download Slides

This talk will focus on how one can build complex data pipelines in Python. I will introduce Luigi and show how it solves problems while running multiple chain of batch jobs like dependency resolution, workflow management, visualisation, failure handling etc.

After that, I will present how to package Luigi pipelines as Docker image for easier testing and deployment. Finally, I will go through way to deploy them on Kubernetes cluster, thus making it possible to scale Big Data pipelines on-demand and reduce infrastructure costs. I will also give tips and tricks to make Luigi Scheduler play well with Kubernetes batch execution feature.

This talk will be accompanied by demo project. It will be very beneficial for audience who have some experience in running batch jobs (not necessarily in Python), typically people who work in Big Data sphere like data scientists, data engineers, BI devs and software developers. Familiarity with Python is helpful but not needed.

Type: Talk (30 mins); Python level: Intermediate; Domain level: Intermediate

Nar Kumar Chhantyal

E. Breuninger GmbH & Co.

Nar Kumar is Data Engineer at where he and his team are responsible for data analytics platform. They use Luigi for data processing and scale the workloads on Google Kubernetes Engine.

Nar Kumar has many years of software development & data engineering experience. He is interested in web services, big data and distributed systems.