airflow(2)
-
[GCP] Airflow + Python (Importing from mysql)
UBUNTU CLIENT CONFIGURATION (리눅스에서 코딩 및 GCP 컨트롤 진행) #gsutil 설치 gcloud auth activate-service-account --key-file= (권한: dataproc editor, storage object creater) or gcloud init export TEMPLATE_ID= export WORK_CLUSTER_NAME= export REGION= export BUCKET_NAME= export PROJECT_ID= export STEP_ID= export PYTHON_FILE=main.py PYTHON CODE $vi pywork.py import pymysql import sys import pandas as pd from datet..
2021.11.11 -
[Dataproc] Pyspark Job with Airrflow (Composer): Get Data From MySQL
UBUNTU CLIENT CONFIGURATION export TEMPLATE_ID=workflow-mytest export WORK_CLUSTER_NAME=cluster-mytest export REGION=asia-northeast3 export BUCKET_NAME=jay-pyspark-mytest #airflow task dag에서도 필요 export PROJECT_ID= #airflow task dag에서도 필요 export PYTHON_FILE=pyspark-job.py export STEP_ID=first_step #Some name like "Get Data" PYTHON CODE $vi pywork.py import pymysql import sys import pandas as pd f..
2021.11.01