dbt-external-tables
dbt macros to stage external sources (by dbt-labs)
dataflow
By HexcloudCo
dbt-external-tables | dataflow | |
---|---|---|
8 | 3 | |
281 | 0 | |
3.9% | - | |
8.4 | 4.5 | |
8 days ago | about 1 year ago | |
PLpgSQL | Shell | |
Apache License 2.0 | MIT License |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
dbt-external-tables
Posts with mentions or reviews of dbt-external-tables.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-04-14.
-
[Need feedback] I wrote a guide about the fundamentals of BigQuery for software developers & traditional database users
You can setup your CDC process such that it will create and evolve the tables for you, e.g. by using jdbc connector. So the moment your OLTP database schema changes, it gets picked up by the CDC process and it will propagate it to your OLAP database. If you want to have more control over schema evolution, you can also do it in dbt using dbt-external-tables package.
-
How to write raw ingestion code when working with dbt?
This looks like what is done by https://github.com/dbt-labs/dbt-external-tables/blob/main/sample_sources/snowflake.yml right ?
-
Personal Project: Beginner seeking advice for Spotify streams project
If you are familiar with dbt, it should be straightforward to do this using dbt and dbt-external-tables.
-
CDC Implementation
We keep track of the schema in git using dbt-external-tables. If a new column is added upstream, whoever made the change is responsible for updating the schema in dbt. It's just a few lines of code. Once PR is merged, the external table that reads the S3 files is refreshed. We never remove columns.
-
What is in your Data Stack? - Thread
In addition to the built-in seed command you can also install this package called dbt-external-tables which will automate the creation of pointers/views to external data sources living on S3/GCS/Azure Blob Storage for several data warehouses. Then you can simply select from these external sources when building your models.
- How to load parquet files from Azure Data Lake Gen2/Azure Blob Storage to Dedicated pool using dbt?
-
Data Pipelines With DBT (Data Build Tool) in Azure
For the external tables, I'm using the following plugin: https://github.com/fishtown-analytics/dbt-external-tables
dataflow
Posts with mentions or reviews of dataflow.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2023-04-14.
-
[Need feedback] I wrote a guide about the fundamentals of BigQuery for software developers & traditional database users
I find these things more useful than clicking around the UI: https://github.com/HexcloudCo/dataflow/blob/main/sinks/gcp/gcp-bigquery-table.sh
-
Need help in solving this architecture
See https://hexcloud.co/docs and https://github.com/HexcloudCo/dataflow/tree/main/sinks/gcp
What are some alternatives?
When comparing dbt-external-tables and dataflow you can also consider the following projects:
dbt-synapse - dbt adapter for Azure Synapse Dedicated SQL Pools
debezium - Change data capture for a variety of databases. Please log issues at https://issues.redhat.com/browse/DBZ.