You need to create a data pipeline that copies time-series transaction data so that it can be queried from within BigQuery by your data science team for analysis.
Every hour, thousands of transactions are updated with a new status. The size of the initial dataset is 1.5 PB, and it will grow by 3 TB per day. The data is heavily structured, and your data science team will build machine learning models based on this data. You want to maximize performance and usability for your data science team. Which two strategies should you adopt? (Choose two.)
rickywck
Highly Voted 3 years, 7 months ago[Removed]
Highly Voted 3 years, 7 months agoawssp12345
2 years, 3 months agomidgoo
Most Recent 7 months, 1 week agovaga1
4 months, 3 weeks agoWillemHendr
4 months, 3 weeks agoodacir
10 months, 3 weeks agozellck
11 months agoAzureDP900
10 months agoNicolasN
12 months agojkhong
10 months, 2 weeks agoNicolasN
12 months agoDerickTW
1 year, 1 month agodevaid
1 year agoMaxNRG
1 year, 9 months agomedeis_jar
1 year, 9 months agodoninakula
1 year, 11 months agosumanshu
2 years, 3 months agoJeysolomon
2 years, 4 months agoChelseajcole
2 years agoretep007
2 years, 1 month agodaghayeghi
2 years, 7 months agoHithesh
2 years, 7 months agosumanshu
2 years, 3 months agoraf2121
2 years, 3 months agohdmi_switch
2 years, 3 months agodaghayeghi
2 years, 8 months agokarthik89
2 years, 8 months agodaghayeghi
2 years, 7 months agoNams_139
2 years, 11 months agofedericohi
2 years, 11 months ago