You have built a model that is trained on data stored in Parquet files. You access the data through a Hive table hosted on Google Cloud. You preprocessed these data with PySpark and exported it as a CSV file into Cloud Storage. After preprocessing, you execute additional steps to train and evaluate your model. You want to parametrize this model training in Kubeflow Pipelines. What should you do?
mil_spyro
Highly Voted 1 year, 11 months agotavva_prudhvi
Highly Voted 1 year, 4 months agomomosoundz
Most Recent 1 year, 4 months agotavva_prudhvi
1 year, 3 months agoM25
1 year, 6 months agoTNT87
1 year, 8 months agochidstar
1 year, 8 months agof084277
1 week, 1 day agoTNT87
1 year, 8 months agoTNT87
1 year, 11 months ago