You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The report generation script currently lives in a crontab on the "worker" VM. We would like to move this into the data-pipeline workflow at the end of the dataflow pipeline.
The text was updated successfully, but these errors were encountered:
It's a little bit trickier as we have three reports:
The bulk of the reports can be run once the old tables are ready
Some reports depend on the httparchive.blink.* tables, which aren't updated until the 1st of the month as they depend on two BigQuery schedule tasks (materialize_blink_features and then Materialize Blink Feature Percentages - which is dependent on the first job). Could they be run as part of the pipeline so we don't have to wait until the first?
The CrUX data is not available until the 2nd Tuesday of the month, so we currently run that on the 15th of the month.
The report generation script currently lives in a crontab on the "worker" VM. We would like to move this into the data-pipeline workflow at the end of the dataflow pipeline.
The text was updated successfully, but these errors were encountered: