Replies: 2 comments
-
Hope this helps! |
Beta Was this translation helpful? Give feedback.
-
|
Now that version 1.0.0 is out I'm hoping AWS will make it available on RDS. If anyone has any insight on that it would be great if they could share here! 🙏 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey, I've been wondering about some questions after coming across this project recently:
Are there any plans on the roadmap for getting the PG_DuckDB extension approved in AWS RDS? We use a managed RDS instance and don't want to set up our own instance or pay some other vendor for a cloud offering.
Can you make a pg_duckdb table that always implicitly reads a Parquet file from S3 when you query it?
Can you have Hive styled partitioning for the aforementioned table, such that you can read multiple Parquet files that back the underlying table/view based on the WHERE filters?
Can you cache said views in order to avoid wasting S3 bandwidth? If yes, how can you invalidate them? Would this be part of a scheduled ETL/ELT pipeline? E.g. periodically ingesting your files in your data lake and then recreating a table/view which stores the underlying Parquet file(s). My main concern is if you have a customer facing frontend which uses DuckDB on the backend to return the data, would the latency (and cost of downloading the entire Parquet file) be a good idea in the long run.
Does the pg_duckdb extension store anything in a
duck.dbfile? If yes, what's it used for, if no, can we treat it as "just" an execution engine?How do transactions work across both the PostgreSQL and DuckDB engine? E.g. can I read from a Parquet file (or a view/table which queries the underlying Parquet file(s), build some analytical query and store the results in a PostgreSQL table?
Beta Was this translation helpful? Give feedback.
All reactions