Question 71

An organization maintains a Google BigQuery dataset that contains tables with user-level data. They want to expose aggregates of this data to other Google Cloud projects, while still controlling access to the user-level data. Additionally, they need to minimize their overall storage cost and ensure the analysis cost for other projects is assigned to those projects. What should they do?
  • Question 72

    You are deploying a new storage system for your mobile application, which is a media streaming service. You decide the best fit is Google Cloud Datastore. You have entities with multiple properties, some of which can take on multiple values. For example, in the entity 'Movie'the property 'actors'and the property 'tags' have multiple values but the property 'date released' does not. A typical query would ask for all movies with actor=<actorname>ordered by date_releasedor all movies with tag=Comedyordered by date_released. How should you avoid a combinatorial explosion in the number of indexes?
  • Question 73

    What are two of the benefits of using denormalized data structures in BigQuery?
  • Question 74

    You have a query that filters a BigQuery table using a WHERE clause on timestamp and ID columns. By using bq query - -dry_run you learn that the query triggers a full scan of the table, even though the filter on timestamp and ID select a tiny fraction of the overall data. You want to reduce the amount of data scanned by BigQuery with minimal changes to existing SQL queries. What should you do?
  • Question 75

    You want to analyze hundreds of thousands of social media posts daily at the lowest cost and with the fewest steps.
    You have the following requirements:
    * You will batch-load the posts once per day and run them through the Cloud Natural Language API.
    * You will extract topics and sentiment from the posts.
    * You must store the raw posts for archiving and reprocessing.
    * You will create dashboards to be shared with people both inside and outside your organization.
    You need to store both the data extracted from the API to perform analysis as well as the raw social media posts for historical archiving. What should you do?