Implement insert_overwrite in s3 #642
mattsmith123
started this conversation in
Ideas
Replies: 1 comment 1 reply
-
Hey @mattsmith123 what you discover it's correct. We first persist the data in a tmp table, then we use it to perform the insert from it. Pretty much https://github.com/dbt-athena/dbt-athena/issues/455 address what you described, feel free to add your thoughts there, and we IMHO should consider to close this discussion and add all in the mentioned issue. And yes, if you are willing to contribute, I'm more than happy to review and make all the necessary steps to add such feature. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
It appears that the current implementation of the
insert_overwrite
incremental strategy is to CTAS a temp table and the do an INSERT INTO from there. This seems to have the unwanted side effect of an extra read and write of the result set. Do I have that wrong? Is there something I don't have configured correctly?Has there been any discussion about a mode that would use S3 to move the data over to the table (basically rename the files) and then just add the partitions? I would maybe be willing to look into implementing it if there is interest and it does not go against the design principles of the project.
Thanks for a great project. This really lets us take advantage of dbt to clean up so really unwieldy queries.
Beta Was this translation helpful? Give feedback.
All reactions