Get hands-on experience with 20+ free Google Cloud products and $300 in free credit for new customers.

BigQuery data extraction

Hello everyone!

I'm running into a problem when extracting data from BigQuery.

My data is 130 GB in total in BigQuery, but I'm being charged for 1 TB of "BigQuery Storage API Network Egress North America to North America". Do you have any idea what might be going on?

Thanks in advance for your attention.

1 3 873
3 REPLIES 3

That sounds like an unexpectedly high cost for data extraction. The billing in BigQuery can sometimes be influenced by the way you're extracting the data. Here are a few potential reasons for this discrepancy:

If you're extracting the data and transferring it to another location or service within North America, you might be incurring egress charges for that transfer. Even if the data extraction itself is within the free tier limits, moving data to another Google Cloud service or to a different location could accrue additional costs.

Inefficient queries that scan unnecessary data or use JOINs extensively can result in higher-than-expected costs. Optimizing your SQL queries to minimize data scanned can reduce costs significantly.

You can try exporting data in a compressed format like gzip can reduce the size of data transferred, potentially decreasing egress charges.

Sometimes, the charges might be due to other activities within your Google Cloud project that generate egress traffic. Ensure that there aren't any other processes or services causing high egress traffic within your project.

To troubleshoot:

1. Double-check the settings and destination of your data extraction process.
2. Review the query used for extraction to ensure it's optimized.
3. Consider using compression and efficient file formats during extraction.
4. Monitor other activities within your project that might be contributing to egress traffic.

For a better solution, reach out to Google Cloud Support with specific details about your extraction process and billing concerns could help in identifying the exact reason behind the unexpected charges.

Hi again @Poala_Tenorio !!

After some research, I find that compressing the data first is not suitable for my case. To give you more context, I'm transferring the data to an environment outside the GCP... I'm extracting the data from the GCP to my on-premise database. Do you think this high cost is due to the fact that I'm moving the data from Google Cloud Platform to my own environment?

Hi @Poala_Tenorio ! Thanks a lot for your response!
I'm gonna explore this 3rd option...
Regarding the other points... (1)Yes we are transferring the data between two different locations: from Mexico to the USA
(2)We are not doing any query, just importing the raw data
(4)We are not doing any other activities within the GCP platform
So I guess this cost is only because of this transaction between locations, what's weird is the 1TB.
But I'll try to compress the data before and see how much will be charged.
Again, many thanks for showing me multiple options to fix this.