Use Hadoop + Hbase on GCP
使用GCP的Cloud Shell來建立叢集

REGION=<region>
CLUSTER_NAME=<cluster_name>
gcloud dataproc clusters create ${CLUSTER_NAME} \
    --optional-components=HBASE,ZOOKEEPER \
    --region=[REGION] \
    --image-version=1.5 \
    --master-machine-type=n1-standard-1 \
    --worker-machine-type=n1-standard-1 \
    --master-boot-disk-size=20GB \
    --worker-boot-disk-size=20GB \
    --enable-component-gateway
加入HUE的版本 port:8888
gcloud beta dataproc clusters create hadoop \
    --optional-components=HBASE,ZOOKEEPER \
    --initialization-actions gs://goog-dataproc-initialization-actions-asia-east1/hue/hue.sh \
    --region=asia-east1 \
    --image-version=1.5 \
    --master-machine-type=n1-standard-1 \
    --worker-machine-type=n1-standard-1 \
    --master-boot-disk-size=20GB \
    --worker-boot-disk-size=20GB \
    --enable-component-gateway
       
但是HUE的設定不會用...
選擇hadoop-m開啟SSH

hdfs dfsadmin -report
顯示hadoop的狀態
sudo jps
GCP要sudo jps才看到得狀態
參考資料:https://github.com/GoogleCloudDataproc/initialization-actions/tree/master/hue
