-
Notifications
You must be signed in to change notification settings - Fork 707
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Airflow BigQueryExampleGen example #54
Comments
Is it possible to use the same pipeline as the tutorial where Airflow & AirflowScheduler are running locally but pull data from BigQuery? Yes How does GCP authentication work in this scenario? It works pretty much the same as kubeflow example, you need to have GOOGLE_APPLICATION_CREDENTIALS setup in your env(in airflow scheduler and webserver console), and set '--project=xxx' in your beam_pipeline_args So basically you can use the taxi_pipeline_simple.py example, with example gen change to BigQueryExampleGen, and beam_pipeline_args setup in additional_pipeline_args To debug, You can try the query in BigQuery web ui console to see if it work, and then try the same query in code |
Got it to work, thanks! |
* Design review for "Attention for Dense Networks"
Hi there,
After going through the workshop tutorial, I am attempting to build my own pipeline ingesting from BigQuery rather than a CSV.
The only example using BigQuery is
taxi_pipeline_kubeflow.py
which assumes execution on GCP.Is it possible to use the same pipeline as the tutorial where Airflow & AirflowScheduler are running locally but pull data from BigQuery?
How does GCP authentication work in this scenario?
I have tried this snippet along with editing
bigquery_default
under admin>connections in the Airflow webapp with no luckThe text was updated successfully, but these errors were encountered: