Live Demo: Build Scalable Event-Driven Microservices with Confluent | Register Now

Presentation

From Raw Data to an Interactive Data App in an Hour: Powered by Snowpark Python

« Current 2023

As data practitioners, we often rely on the data engineering teams upstream to deliver the right data needed to train ML models at scale. Deploying these ML models as a data application to downstream business users is constrained by one’s web development experience. Using Snowpark, you can build end to end data pipelines, and data applications from scratch using Python.

In this talk, you will learn to build a Streamlit data application to help visualize the ROI of different advertising spends of an example organization.

  • Setup Environment: Use stages and tables to ingest and organize raw data from S3 into Snowflake.
  • Data Engineering: Leverage Snowpark for Python DataFrames to perform data transformations such as group by, aggregate, pivot, and join to prep the data for downstream applications.
  • Data Pipelines: Use Snowflake Tasks to turn your data pipeline code into operational pipelines with integrated monitoring.
  • Machine Learning: Prepare data and run ML Training in Snowflake using Snowpark ML and deploy the model as a Snowpark User-Defined-Function (UDF).
  • Streamlit Application: Build an interactive application using Python (no web development experience required) to help visualize the ROI of different advertising spend budgets.

Related Links

How Confluent Completes Apache Kafka eBook

Leverage a cloud-native service 10x better than Apache Kafka

Confluent Developer Center

Spend less on Kafka with Confluent, come see how