You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The BQ tables are getting unwieldy and expensive to query. We use scheduled queries to generate the tables in the latest dataset. Similarly, we should generate a subset of these tables randomly limited to some number of rows to ensure that the query will be inexpensive. For the requests dataset we should group by page so that all pages in the sample have all of their respective requests.
calculate the average row size in bytes for each dataset
pick a sample size in rows corresponding to about 1 GB per dataset (this can change)
schedule a query for each dataset (requests, pages, etc) and each client (desktop, mobile) to materialize sample tables
The text was updated successfully, but these errors were encountered:
The BQ tables are getting unwieldy and expensive to query. We use scheduled queries to generate the tables in the
latest
dataset. Similarly, we should generate a subset of these tables randomly limited to some number of rows to ensure that the query will be inexpensive. For therequests
dataset we should group bypage
so that all pages in the sample have all of their respective requests.The text was updated successfully, but these errors were encountered: