Question 76

You are building a model to make clothing recommendations. You know a user's fashion pis likely to change over time, so you build a data pipeline to stream new data back to the model as it becomes available. How should you use this data to train the model?
  • Question 77

    You designed a database for patient records as a pilot project to cover a few hundred patients in three clinics. Your design used a single database table to represent all patients and their visits, and you used self-joins to generate reports. The server resource utilization was at 50%. Since then, the scope of the project has expanded. The database must now store 100 times more patient records. You can no longer run the reports, because they either take too long or they encounter errors with insufficient compute resources. How should you adjust the database design?
  • Question 78

    You are building a new application that you need to collect data from in a scalable way. Data arrives continuously from the application throughout the day, and you expect to generate approximately 150 GB of JSON data per day by the end of the year. Your requirements are:
    * Decoupling producer from consumer
    * Space and cost-efficient storage of the raw ingested data, which is to be stored indefinitely
    * Near real-time SQL query
    * Maintain at least 2 years of historical data, which will be queried with SQ Which pipeline should you use to meet these requirements?
  • Question 79

    You need to create a data pipeline that copies time-series transaction data so that it can be queried from within BigQuery by your data science team for analysis. Every hour, thousands of transactions are updated with a new status. The size of the intitial dataset is 1.5 PB, and it will grow by 3 TB per day. The data is heavily structured, and your data science team will build machine learning models based on this data. You want to maximize performance and usability for your data science team. Which two strategies should you adopt?
    (Choose two.)
  • Question 80

    Your company is performing data preprocessing for a learning algorithm in Google Cloud Dataflow.
    Numerous data logs are being are being generated during this step, and the team wants to analyze them.
    Due to the dynamic nature of the campaign, the data is growing exponentially every hour. The data scientists have written the following code to read the data for a new key features in the logs.
    BigQueryIO.Read
    .named("ReadLogData")
    .from("clouddataflow-readonly:samples.log_data")
    You want to improve the performance of this data read. What should you do?