We haven't got any application logs in bigquery since the weekend's auto update.
kube-system is still getting there so routing is OK - looks like It's doing "system" but not "workload"?
Any ideas? Downgrade?
Solved! Go to Solution.
So checking the logs helped.
The version update was coincidental - although it might have introduced some changes to some of the log messages which is what prompted the real problem which is that any variatrion in the log messages causes the creation of new columns. (One of ingresses logs http headers and that includes anything that the client might have in them). Over time these build up and exceeed the 10K column limit.
The lack of control over the schema is a big problem with bigquery as a logging sink. It already floods us with noise emails about incompatible schema when two components log incompatible things. That in turn makes it harder to see the real problems as you get into habit of ignoring the mails from logging.
So the actual fix was just to delete all of the historical logs and start again.
In the medium term we'll probably get off bigquery as the schema is a all a bit of kludge and use something that we can actually configure.
Hi @Tim_Murnaghan,
Welcome to Google Cloud Community!
This page provides production updates for Google Kubernetes Engine (GKE), including announcements on updated features, bug fixes, known issues, and deprecated functionality. It covers release notes for all channels and versions, allowing you to stay informed about the latest changes.
To address your workload logging issues, consider the following steps:
- Pod Logs: Check the pod logs in your application namespaces if there are any errors related to logging.
- Logging Driver: Ensure the logging driver specified in your pods' containers is correctly configured.
By reviewing workload-specific logging configurations and troubleshooting steps, you can identify why workload logs are not reaching BigQuery. If the issue persists, contact Google Cloud Support with detailed logs, configurations, and error messages for further assistance.
Was this helpful? If so, please accept this answer as “Solution”. If you need additional assistance, reply here within 2 business days and I’ll be happy to help.
The sudden halt of application logs flowing into BigQuery after a GKE auto-update strongly suggests a compatibility issue or a configuration change introduced by the update. Let's troubleshoot this systematically:
1. Confirm Logging Agent Status
2. Verify BigQuery Logging Configuration
3. GKE Update Impact
4. Filtering Issues
5. Potential Solutions
Troubleshooting Steps
By following these steps, you should be able to identify the cause of the problem and restore your application logs to BigQuery.
So checking the logs helped.
The version update was coincidental - although it might have introduced some changes to some of the log messages which is what prompted the real problem which is that any variatrion in the log messages causes the creation of new columns. (One of ingresses logs http headers and that includes anything that the client might have in them). Over time these build up and exceeed the 10K column limit.
The lack of control over the schema is a big problem with bigquery as a logging sink. It already floods us with noise emails about incompatible schema when two components log incompatible things. That in turn makes it harder to see the real problems as you get into habit of ignoring the mails from logging.
So the actual fix was just to delete all of the historical logs and start again.
In the medium term we'll probably get off bigquery as the schema is a all a bit of kludge and use something that we can actually configure.