Skip to content

Commit

Permalink
fix s3 integ tests (#1003)
Browse files Browse the repository at this point in the history
  • Loading branch information
surjits254 authored Oct 22, 2024
1 parent 8c5cf60 commit 0d617bd
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 4 deletions.
2 changes: 1 addition & 1 deletion python/.ci/Jenkinsfile
Original file line number Diff line number Diff line change
Expand Up @@ -595,7 +595,7 @@ pipeline {
--s3.bq.output.dataset.name="dataproc_templates" \
--s3.bq.output.table.name="s3_to_bq_avro_py" \
--s3.bq.output.mode="overwrite" \
--s3.bq.temp.bucket.name="${GCS_STAGING_LOCATION}"
--s3.bq.temp.bucket.name="python-dataproc-templates-temp"
'''
}
}
Expand Down
3 changes: 0 additions & 3 deletions python/dataproc_templates/s3/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,6 @@

Template for reading files from Amazon S3 and writing them to a BigQuery table. It supports reading JSON, CSV, Parquet and Avro formats.

It uses the [Spark BigQuery connector](https://cloud.google.com/dataproc-serverless/docs/guides/bigquery-connector-spark-example) for writing to BigQuery.

## Arguments

* `s3.bq.input.location` : Amazon S3 input location. Input location must begin with `s3a://`
Expand Down Expand Up @@ -60,7 +58,6 @@ export GCS_STAGING_LOCATION=<gcs-bucket-name>
export GCP_PROJECT=my-project
export REGION=us-west1
export GCS_STAGING_LOCATION=my-staging-bucket
export JARS=gs://spark-lib/bigquery/spark-bigquery-latest_2.12.jar

./bin/start.sh \
-- --template=S3TOBIGQUERY \
Expand Down

0 comments on commit 0d617bd

Please sign in to comment.