Question 101

When you design a Google Cloud Bigtable schema it is recommended that you
_________.
  • Question 102

    You are implementing security best practices on your data pipeline. Currently, you are manually executing
    jobs as the Project Owner. You want to automate these jobs by taking nightly batch files containing non-
    public information from Google Cloud Storage, processing them with a Spark Scala job on a Google Cloud
    Dataproc cluster, and depositing the results into Google BigQuery.
    How should you securely run this workload?
  • Question 103

    Cloud Dataproc charges you only for what you really use with _____ billing.
  • Question 104

    Government regulations in your industry mandate that you have to maintain an auditable record of access
    to certain types of data. Assuming that all expiring logs will be archived correctly, where should you store
    data that is subject to that mandate?
  • Question 105

    You are creating a new pipeline in Google Cloud to stream IoT data from Cloud Pub/Sub through Cloud Dataflow to BigQuery. While previewing the data, you notice that roughly 2% of the data appears to be corrupt.
    You need to modify the Cloud Dataflow pipeline to filter out this corrupt data. What should you do?