Optimizing Usage of Cloud Data Loss Prevention (DLP) API for Cost Reduction

Cost Reduction Options for Cloud Data Loss Prevention (DLP) API

Question

As adoption of the Cloud Data Loss Prevention (DLP) API grows within the company, you need to optimize usage to reduce cost.

DLP target data is stored in Cloud Storage and BigQuery.

The location and region are identified as a suffix in the resource name.

Which cost reduction options should you recommend?

Answers

Explanations

Click on the arrows to vote for the correct answer

A. B. C. D.

C.

https://cloud.google.com/dlp/docs/reference/rest/v2/InspectJobConfig

The Cloud Data Loss Prevention (DLP) API is a tool that helps organizations identify and protect sensitive data across their cloud-based infrastructure. As adoption of the DLP API grows within a company, it is important to optimize its usage to reduce cost.

In this scenario, the DLP target data is stored in Cloud Storage and BigQuery, and the location and region are identified as a suffix in the resource name. To reduce cost, the following cost reduction options should be recommended:

Option A:

  • Set appropriate rowsLimit value on BigQuery data hosted outside the US: This option suggests limiting the number of rows scanned during a DLP operation on BigQuery data hosted outside the US. By setting an appropriate rowsLimit value, the DLP API will only scan the specified number of rows, thereby reducing the amount of data scanned and lowering costs.
  • Set appropriate bytesLimitPerFile value on multiregional Cloud Storage buckets: This option suggests setting an appropriate bytesLimitPerFile value on multiregional Cloud Storage buckets to limit the amount of data scanned during a DLP operation. By setting a limit on the amount of data scanned, the DLP API will only scan the specified amount of data, thereby reducing costs.

Option B:

  • Set appropriate rowsLimit value on BigQuery data hosted outside the US: This option is similar to Option A, where the number of rows scanned during a DLP operation on BigQuery data hosted outside the US is limited.
  • Minimize transformation units on multiregional Cloud Storage buckets: This option suggests minimizing the number of transformation units used on multiregional Cloud Storage buckets to reduce the amount of data scanned during a DLP operation. By minimizing transformation units, the DLP API will only scan the specified amount of data, thereby reducing costs.

Option C:

  • Use rowsLimit and bytesLimitPerFile to sample data: This option suggests using rowsLimit and bytesLimitPerFile to sample data during a DLP operation. By sampling data, the DLP API will only scan a subset of the data, thereby reducing costs.
  • Use CloudStorageRegexFileSet to limit scans: This option suggests using CloudStorageRegexFileSet to limit the scans to only specific files in Cloud Storage. By limiting the scans to specific files, the DLP API will only scan the specified files, thereby reducing costs.

Option D:

  • Use FindingLimits and TimespanConfig to sample data: This option suggests using FindingLimits and TimespanConfig to sample data during a DLP operation. By sampling data, the DLP API will only scan a subset of the data, thereby reducing costs.
  • Minimize transformation units: This option suggests minimizing the number of transformation units used to reduce the amount of data scanned during a DLP operation. By minimizing transformation units, the DLP API will only scan the specified amount of data, thereby reducing costs.

In summary, all of the options presented offer potential cost reduction strategies for optimizing usage of the Cloud Data Loss Prevention (DLP) API. However, the most appropriate option(s) will depend on the specific needs and resources of the company in question.