-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Spike: Prove Fast API with DuckDB and Parquet on S3 #102
Comments
Code along with README has been pushed to the new pipeline-internal-api repository: https://github.com/digital-land/pipeline-internal-api |
The work completed in this spike certainly contributed to the implementation required for #106. However, it doesn't provide the testing and deployment into AWS which will be required for ticket 106. |
Reviewed with the team and made one change: Switched over from explicit path manipulation for dataset and resource to replying upon the automatic hive_partitioning inference built into DuckDB. Dataset and resource are now just WHERE clause parameters, like the other query parameters |
Overview
Following the design proposal for an internal API, we would like to prove some technology choices which include the use of Fast API with DuckDB accessing Parquet on S3.
This work was identified during the spike on API design.
Tech Approach
Acceptance Criteria/Tests
Ticket Management - DELETE this section once completed
The text was updated successfully, but these errors were encountered: