How to use Kedro for Lakehouse ETL ? #1009
Replies: 1 comment 7 replies
-
Hi @nthanhtung you ask a very good question - something which truly makes us ask what Kedro should and shouldn't be. As the product has evolved, my view is that it is best described as a 'ML engineering framework'. What does this mean in practice? Kedro provides a mechanism for organising your Python ML process in a structured, battle-tested pattern in a data-centric way of working with batch data. Going through your requirements, I've gone through how I think Kedro fits into the world you describe:
To summarise I think point 11 is the most important here - there is a lot of complexity in the middle part of that process that at a macro level is just 'a part of the ETL process', but something Kedro tries to focus on. A mature system could look something like this:
|
Beta Was this translation helpful? Give feedback.
-
How to use Kedro for Lakehouse ETL ?
Lakehouse Architecture
https://www.youtube.com/watch?v=d7Px_MZV05I
Requirement for ETL Tool
Easy for traditional database developer, citizen developer to adopt
Available ETL Tool that can support the above requirement
Azure Data Factory
Databricks Delta
Beta Was this translation helpful? Give feedback.
All reactions