You need to migrate Hadoop jobs for your company's Data Science team without modifying the underlying infrastructure.
You want to minimize costs and infrastructure management effort.
What should you do?
Click on the arrows to vote for the correct answer
A. B. C. D.A.
https://cloud.google.com/architecture/hadoop/hadoop-gcp-migration-jobsTo migrate Hadoop jobs for your company's Data Science team without modifying the underlying infrastructure and minimize costs and infrastructure management effort, you should choose option B: Create a Dataproc cluster using preemptible worker instances.
Dataproc is a fully-managed cloud service that lets you easily run Apache Hadoop and Apache Spark jobs. With Dataproc, you can create a Hadoop cluster quickly and easily, without having to manually deploy and configure the underlying infrastructure. Dataproc provides a number of features that help you optimize the performance and cost of your Hadoop jobs, including auto-scaling, cluster management, and monitoring.
Preemptible worker instances are a cost-effective option for running Hadoop jobs in Dataproc. These instances are the same as regular Compute Engine instances, but with the caveat that Google may reclaim them at any time with a 30-second warning. Because preemptible instances are priced at a significant discount compared to regular instances, they can help you save money on your Hadoop jobs.
Here are some additional reasons why choosing option B is the best choice:
In summary, creating a Dataproc cluster using preemptible worker instances is the most cost-effective and efficient option for migrating Hadoop jobs for your company's Data Science team.