Environment Forking for Flexible Data Source Configuration #4298
Replies: 4 comments
-
Hi @pascalwhoop, thanks for bringing these suggestion, they seem worthwhile exploring! |
Beta Was this translation helpful? Give feedback.
-
Turning this into a discussion 🙏🏼 Let's continue the conversation there. |
Beta Was this translation helpful? Give feedback.
-
About |
Beta Was this translation helpful? Give feedback.
-
About |
Beta Was this translation helpful? Give feedback.
-
We're considering adding a feature to allow more flexible configuration of data sources across environments. The primary use case is to enable testing part of the pipeline using production data without needing to copy data manually. Thought I'd share here to see if others find this useful as well.
Proposed Features:
Environment Forking Flag:
Example:
kedro run --from-nodes a,b,c --fork-from prod --env dev
This would read initial datasets from the 'prod' environment and then execute the rest of the pipeline in the 'dev' environment.
Dataset Copying Command:
Example:
kedro copy --datasets a,b,c --from prod --to dev
This would manually copy specified datasets from 'prod' to 'dev' environment before running the pipeline.
Inverse Tag Filtering:
Example:
kedro run --without-tags tag1,tag2
This would filter out nodes based on tags, inverse of the existing
--tags
option.Use Case:
Current Limitations:
Potential Implementation:
Long-term Consideration:
Beta Was this translation helpful? Give feedback.
All reactions