You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have searched the existing issues, and I could not find an existing issue for this bug
Current Behavior
I have a model that writes out to more than 100 hive partitions.
I recently dropped the table and attempted to recreate via dbt build. I have an external_location in configuration as I was migrated onto dbt and want to avoid rebuild data / move s3 location.
The unexpected behavior is dbt build succeeded but my s3 path does not contain any data.
On model.A.XX: -- /* {"app": "dbt", "dbt_version": "1.7.17", "profile_name": "A", "target_name": "prod", "node_id": "model.A.XX"} */
create table "awsdatacatalog"."{schema}"."XX__tmp_not_partitioned"
with (
table_type='hive',
is_external=true,external_location= {config.external_location}
format='orc'
)
as
SELECT
Expected Behavior
This XX__tmp_not_partitioned should not use the configured external path which is clashing with the actual model. Instead it should either use a unique tmp location or same as __dbt_tmp table which goes to s3_data_dir/schema/table__dbt_tmp
Steps To Reproduce
create a partitioned model with more than 100 hive partitions, with an external_location in config.
@lucastrubiano
I recommend to have a look at the contributing section.
Long story short, if you are not a project maintainer you cannot push directly to dbt-athena/dbt-athena.
You have to fork the repository, and then propose a PR from your fork, pretty much here you can find a detailed guide on how to do it.
Is this a new bug in dbt-athena?
Current Behavior
I have a model that writes out to more than 100 hive partitions.
I recently dropped the table and attempted to recreate via dbt build. I have an external_location in configuration as I was migrated onto dbt and want to avoid rebuild data / move s3 location.
The unexpected behavior is dbt build succeeded but my s3 path does not contain any data.
dbt / adaptor version:
The CTAS generated:
Expected Behavior
This XX__tmp_not_partitioned should not use the configured external path which is clashing with the actual model. Instead it should either use a unique tmp location or same as __dbt_tmp table which goes to s3_data_dir/schema/table__dbt_tmp
Steps To Reproduce
create a partitioned model with more than 100 hive partitions, with an external_location in config.
If this table exists in glue, drop it.
Run dbt run to create more than 100 partitions in one sql run.
Environment
Additional Context
I am happy to contribute to get this fixed if that helps speed things up.
https://github.com/dbt-athena/dbt-athena/blob/main/dbt/include/athena/macros/materializations/models/table/create_table_as.sql
The text was updated successfully, but these errors were encountered: