What should you do?

Your company’s on-premises Apache Hadoop servers are approaching end-of-life, and IT has decided to migrate the cluster to Google Cloud Dataproc. A likefor- like migration of the cluster would require 50 TB of Google Persistent Disk per node. The CIO is concerned about the cost of using that much block storage.
You want to minimize the storage cost of the migration. What should you do?
A. Put the data into Google Cloud Storage.
B. Use preemptible virtual machines (VMs) for the Cloud Dataproc cluster.
C. Tune the Cloud Dataproc cluster so that there is just enough disk for all data.
D. Migrate some of the cold data into Google Cloud Storage, and keep only the hot data in Persistent Disk.

Download Printable PDF. VALID exam to help you PASS.

2 thoughts on “What should you do?

  1. A. This is Google recommandation to move persistant data to GC STorage. (https://cloud.google.com/solutions/migration/hadoop/hadoop-gcp-migration-overview)
    B. You should not use preemptible VM for persistent storage.
    C. It is not recommanded to have just enough disk if the company plan to have more data in the future.
    D. Migrating fully to GCS can increase IO variance and latency (https://cloud.google.com/blog/products/storage-data-transfer/hdfs-vs-cloud-storage-pros-cons-and-migration-tips), which can lead this company to have a more hybrid approach using GPD which guarantees high IOPS.

Leave a Reply

Your email address will not be published. Required fields are marked *


The reCAPTCHA verification period has expired. Please reload the page.