Use Hadoop + Hbase on GCP

GCP 2020年6月21日

使用GCP的Cloud Shell來建立叢集

Cloud Shell

REGION=<region>
CLUSTER_NAME=<cluster_name>
gcloud dataproc clusters create ${CLUSTER_NAME} \
    --optional-components=HBASE,ZOOKEEPER \
    --region=[REGION] \
    --image-version=1.5 \
    --master-machine-type=n1-standard-1 \
    --worker-machine-type=n1-standard-1 \
    --master-boot-disk-size=20GB \
    --worker-boot-disk-size=20GB \
    --enable-component-gateway

加入HUE的版本 port:8888

gcloud beta dataproc clusters create hadoop \
    --optional-components=HBASE,ZOOKEEPER \
    --initialization-actions gs://goog-dataproc-initialization-actions-asia-east1/hue/hue.sh \
    --region=asia-east1 \
    --image-version=1.5 \
    --master-machine-type=n1-standard-1 \
    --worker-machine-type=n1-standard-1 \
    --master-boot-disk-size=20GB \
    --worker-boot-disk-size=20GB \
    --enable-component-gateway
       

但是HUE的設定不會用...

選擇hadoop-m開啟SSH
selected hadoop-m

hdfs dfsadmin -report

顯示hadoop的狀態

sudo jps

GCP要sudo jps才看到得狀態

參考資料:https://github.com/GoogleCloudDataproc/initialization-actions/tree/master/hue

sheepdragon

五樓的樓下是幾樓呢?

Great! You've successfully subscribed.
Great! Next, complete checkout for full access.
Welcome back! You've successfully signed in.
Success! Your account is fully activated, you now have access to all content.