Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Professional Data Engineer All Questions

View all questions & answers for the Professional Data Engineer exam

Exam Professional Data Engineer topic 1 question 269 discussion

Actual exam question from Google's Professional Data Engineer
Question #: 269
Topic #: 1
[All Professional Data Engineer Questions]

Your organization's data assets are stored in BigQuery, Pub/Sub, and a PostgreSQL instance running on Compute Engine. Because there are multiple domains and diverse teams using the data, teams in your organization are unable to discover existing data assets. You need to design a solution to improve data discoverability while keeping development and configuration efforts to a minimum. What should you do?

  • A. Use Data Catalog to automatically catalog BigQuery datasets. Use Data Catalog APIs to manually catalog Pub/Sub topics and PostgreSQL tables.
  • B. Use Data Catalog to automatically catalog BigQuery datasets and Pub/Sub topics. Use Data Catalog APIs to manually catalog PostgreSQL tables.
  • C. Use Data Catalog to automatically catalog BigQuery datasets and Pub/Sub topics. Use custom connectors to manually catalog PostgreSQL tables.
  • D. Use customer connectors to manually catalog BigQuery datasets, Pub/Sub topics, and PostgreSQL tables.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?) , you can switch to a simple comment.
Switch to a voting comment New
datapassionate
Highly Voted 10 months ago
Selected Answer: C
Data Catalog is the best choice. But for catalogging PostgreSQL it is better to use a connector when available, instead of using API. https://cloud.google.com/data-catalog/docs/integrate-data-sources#integrate_unsupported_data_sources
upvoted 11 times
7787de3
2 months, 1 week ago
I agree. On the linked page: If you can't find a connector for your data source, you can still manually integrate it by creating entry groups and custom entries. As we can find a connector there, it should be used.
upvoted 1 times
...
tibuenoc
9 months, 3 weeks ago
Agree. If it doesn't have a connector, it must be manually built on the Data Catalog API. As PostgreSQL already has a connector it's the best option is C
upvoted 3 times
...
...
raaad
Highly Voted 10 months, 3 weeks ago
Selected Answer: B
- It utilizes Data Catalog's native support for both BigQuery datasets and Pub/Sub topics. - For PostgreSQL tables running on a Compute Engine instance, you'd use Data Catalog APIs to create custom entries, as Data Catalog does not automatically discover external databases like PostgreSQL.
upvoted 11 times
AllenChen123
10 months, 1 week ago
Agree. https://cloud.google.com/data-catalog/docs/concepts/overview#catalog-non-google-cloud-assets
upvoted 4 times
...
...
baimus
Most Recent 1 month, 2 weeks ago
Selected Answer: C
This is C. To clarify some issues below with B, the links provided by supporters of B actually do say that it's preferable to use a community connector where available, and to only use the API when the case is genuinely not supported by community connectors. In this case it's Postgresql, so it's supported, see here for full list: https://cloud.google.com/data-catalog/docs/integrate-data-sources#integrate_on-premises_data_sources So this would be B if it was something like Q+ or some genuinely unsupported database, but postgres is supported for community connector.
upvoted 1 times
...
shanks_t
3 months ago
Selected Answer: B
Data Catalog automatically catalogs metadata from Google Cloud sources such as BigQuery, Vertex AI, Pub/Sub, Spanner, Bigtable, and more. To catalog metadata from non-Google Cloud systems in your organization, you can use the following: Community-contributed connectors to multiple popular on-premises data sources Manually build on the Data Catalog APIs for custom entries
upvoted 2 times
shanks_t
3 months ago
C. While similar to B, using custom connectors for PostgreSQL might involve more development effort than using the Data Catalog APIs directly.
upvoted 1 times
...
...
meh_33
3 months, 2 weeks ago
raaad mostly correct and we can check his description supporting his answer so we can go with it .Cheers mate
upvoted 1 times
...
987af6b
4 months ago
Selected Answer: C
I’m voting for C because the documentation states that Postgres is a custom connector developed by the community.
upvoted 1 times
987af6b
3 months, 4 weeks ago
Changed my mind. B. -This is not on premise, so the custom connector should not be applicable -Question says keep manual dev and config to a minimum
upvoted 1 times
...
...
fitri001
5 months, 1 week ago
Selected Answer: B
BigQuery Datasets and Pub/Sub Topics: Google Data Catalog can automatically catalog metadata from BigQuery and Pub/Sub, making it easy to discover and manage these data assets without additional development effort. PostgreSQL Tables: While Data Catalog does not have built-in connectors for PostgreSQL, you can use the Data Catalog APIs to manually catalog the PostgreSQL tables. This requires some custom development but is manageable compared to creating custom connectors for everything.
upvoted 2 times
...
virat_kohli
6 months ago
Selected Answer: B
B. Use Data Catalog to automatically catalog BigQuery datasets and Pub/Sub topics. Use Data Catalog APIs to manually catalog PostgreSQL tables.
upvoted 1 times
...
Cassim
6 months, 2 weeks ago
Selected Answer: B
Option B leverages Data Catalog to automatically catalog BigQuery datasets and Pub/Sub topics, which streamlines the process and reduces manual effort. Using Data Catalog APIs to manually catalog PostgreSQL tables ensures consistency across all data assets while minimizing development and configuration efforts.
upvoted 1 times
...
LaxmanTiwari
7 months ago
Selected Answer: C
I vote for c as per Integrate on-premises data sources To integrate on-premises data sources, you can use the corresponding Python connectors contributed by the community: under the link https://cloud.google.com/data-catalog/docs/integrate-data-sources
upvoted 2 times
LaxmanTiwari
7 months ago
data catalog api will come into effect if custom connectors are not available via community repos.
upvoted 1 times
...
...
joao_01
7 months, 3 weeks ago
In the opction C, the expression "Use custom connectors to manually catalog PostgreSQL tables." is refering to the use case of Google when you want to use "Community-contributed connectors to multiple popular on-premises data sources". As you can see, this connectors are for ON-PREMISSES data sources ONLY. In this case the Postgres is in a VM in the cloud. Thus, the option correct is B.
upvoted 3 times
joao_01
7 months, 3 weeks ago
Link: https://cloud.google.com/data-catalog/docs/concepts/overview#catalog-non-google-cloud-assets
upvoted 1 times
...
...
hanoverquay
8 months, 1 week ago
Selected Answer: B
option B, there's no need to build a custom connector now, postgreSQL is now supported https://github.com/GoogleCloudPlatform/datacatalog-connectors-rdbms/tree/master/google-datacatalog-postgresql-connector
upvoted 1 times
d11379b
8 months ago
I think “custom connector” here may just infer that this is not official tools? as the doc mentioned “ connectors contributed by the community” And should not be B as “manually catalog by API “ this is a way even more basic than using connector
upvoted 1 times
...
...
Y___ash
8 months, 2 weeks ago
Selected Answer: B
Use Data Catalog to automatically catalog BigQuery datasets and Pub/Sub topics. Use Data Catalog APIs to manually catalog PostgreSQL tables.
upvoted 1 times
...
Harshzh12
9 months ago
Selected Answer: B
Datacatalog API contain the connector for postgresql with using it developer don't have to create the custom connectors
upvoted 1 times
...
ML6
9 months, 1 week ago
Selected Answer: C
Google Recommendation: If you can't find a connector for your data source, you can still manually integrate it by creating entry groups and custom entries. To do that, you can: - Use one of the Data Catalog Client Libraries in one of the following languages: C#, Go, Java, Node.js, PHP, Python, or Ruby. - Or manually build on the Data Catalog API. However, there is a connector for PostgreSQL, so option C.
upvoted 4 times
...
saschak94
9 months, 2 weeks ago
Selected Answer: C
If you can't find a connector for your data source, you can still manually integrate it by creating entry groups and custom entries. To do that, you can: - Manually build on the Data Catalog API.
upvoted 2 times
...
Matt_108
10 months, 2 weeks ago
Selected Answer: B
Option B - Data Catalog automatically maps out GCP resources and dev efforts are minimized by leveraging the data catalog API to do the same for postgresql db
upvoted 3 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...