-
Notifications
You must be signed in to change notification settings - Fork 0
/
ETL_toll_data.py
74 lines (65 loc) · 2.26 KB
/
ETL_toll_data.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
from datetime import timedelta
from airflow import DAG
from airflow.operators.bash_operator import BashOperator
from airflow.utils.dates import days_ago
# Task 1.1 - Define DAG arguments
default_args = {
'owner': 'Adam',
'start_date': days_ago(0),
'email': ['myemail@mail.com'],
'email_on_failure': True,
'email_on_retry': True,
'retries': 1,
'retry_delay': timedelta(minutes=5)
}
# Task 1.2 - Define the DAG
dag = DAG(
'ETL_toll_data',
default_args=default_args,
schedule_interval=timedelta(days=1),
description='Apache Airflow Final Assignment',
)
# Task 1.3 - Create a task to unzip data
unzip_data = BashOperator(
task_id='unzip_data',
bash_command='tar -xvzf /home/project/airflow/dags/finalassignment/tolldata.tgz',
dag=dag,
cwd=dag.folder,
)
# Task 1.4 - Create a task to extract data from csv file
extract_data_from_csv = BashOperator(
task_id='extract_data_from_csv',
bash_command='cut -d"," -f1-4 vehicle-data.csv > csv_data.csv',
dag=dag,
cwd=dag.folder,
)
# Task 1.5 - Create a task to extract data from tsv file
extract_data_from_tsv = BashOperator(
task_id='extract_data_from_tsv',
bash_command='cut -f5-7 tollplaza-data.tsv | tr "\t" "," | tr -d "\r" > tsv_data.csv',
dag=dag,
cwd=dag.folder,
)
# Task 1.6 - Create a task to extract data from fixed width file
extract_data_from_fixed_width = BashOperator(
task_id='extract_data_from_fixed_width',
bash_command='cut -c59-67 payment-data.txt | tr " " "," > fixed_width_data.csv',
dag=dag,
cwd=dag.folder,
)
# Task 1.7 - Create a task to consolidate data extracted from previous tasks
consolidate_data = BashOperator(
task_id='consolidate_data',
bash_command='paste -d"," csv_data.csv tsv_data.csv fixed_width_data.csv > extracted_data.csv',
dag=dag,
cwd=dag.folder,
)
# Task 1.8 - Transform and load the data
transform_data = BashOperator(
task_id='transform_data',
bash_command='awk -F\',\' \'{print $1","$2","$3","toupper($4)","$5","$6","$7","$8","$9}\' < extracted_data.csv > transformed_data.csv',
dag=dag,
cwd=dag.folder,
)
# Task 1.9 - Define the task pipeline
unzip_data >> extract_data_from_csv >> extract_data_from_tsv >> extract_data_from_fixed_width >> consolidate_data >> transform_data