15 Data pipeline continuous integration
This chapter covers
- Separating the data engineering environments
- Database change management
- Configuring Snowflake to use Git
- Using the Snowflake CLI command line interface
- Connecting to Snowflake securely
In previous chapters, we gradually built data pipelines by adding various pieces of functionality. As our knowledge expanded, we created many scripts and files, saving them across multiple chapter folders in the accompanying GitHub repository. This has made it challenging to locate a specific script for maintenance. A more practical solution for organizing the data pipeline code would be to store the scripts in an organized manner in the repository. A centralized code repository is essential when multiple data engineers work on the same data pipelines, allowing them to locate scripts effortlessly and merge their code changes into the shared codebase.