Saving data using Spark BigQuery connector with ranged index
xtrmstep opened this issue · 5 comments
Hi,
Could you, please, help with the described problem related to the connector? If this is possible at all.
The problem described in the question on stackoverflow: https://stackoverflow.com/questions/75142769/saving-data-using-spark-bigquery-connector-with-ranged-index
Added support for this.
@vishalkarve15 could you, please, advise about the documentation link to understand how exactly is possible to use it?
It will be present in the next release. For now, you can look at the partitionField
option in https://github.com/GoogleCloudDataproc/spark-bigquery-connector/blob/master/README-template.md
If you want to use it right away, you can build the latest commit.
You can also try one of the nightly builds (based on your spark version):
- gs://spark-lib-nightly-snapshots/spark-2.4-bigquery-0.0.20230808.jar
- gs://spark-lib-nightly-snapshots/spark-3.1-bigquery-0.0.20230808.jar
- gs://spark-lib-nightly-snapshots/spark-3.2-bigquery-0.0.20230808.jar
- gs://spark-lib-nightly-snapshots/spark-3.3-bigquery-0.0.20230808.jar
- gs://spark-lib-nightly-snapshots/spark-3.4-bigquery-0.0.20230808.jar