exam questions

Exam AWS Certified Solutions Architect - Professional SAP-C02 All Questions

View all questions & answers for the AWS Certified Solutions Architect - Professional SAP-C02 exam

Exam AWS Certified Solutions Architect - Professional SAP-C02 topic 1 question 147 discussion

A financial services company receives a regular data feed from its credit card servicing partner. Approximately 5,000 records are sent every 15 minutes in plaintext, delivered over HTTPS directly into an Amazon S3 bucket with server-side encryption. This feed contains sensitive credit card primary account number (PAN) data. The company needs to automatically mask the PAN before sending the data to another S3 bucket for additional internal processing. The company also needs to remove and merge specific fields, and then transform the record into JSON format. Additionally, extra feeds are likely to be added in the future, so any design needs to be easily expandable.

Which solutions will meet these requirements?

  • A. Invoke an AWS Lambda function on file delivery that extracts each record and writes it to an Amazon SQS queue. Invoke another Lambda function when new messages arrive in the SQS queue to process the records, writing the results to a temporary location in Amazon S3. Invoke a final Lambda function once the SQS queue is empty to transform the records into JSON format and send the results to another S3 bucket for internal processing.
  • B. Invoke an AWS Lambda function on file delivery that extracts each record and writes it to an Amazon SQS queue. Configure an AWS Fargate container application to automatically scale to a single instance when the SQS queue contains messages. Have the application process each record, and transform the record into JSON format. When the queue is empty, send the results to another S3 bucket for internal processing and scale down the AWS Fargate instance.
  • C. Create an AWS Glue crawler and custom classifier based on the data feed formats and build a table definition to match. Invoke an AWS Lambda function on file delivery to start an AWS Glue ETL job to transform the entire record according to the processing and transformation requirements. Define the output format as JSON. Once complete, have the ETL job send the results to another S3 bucket for internal processing.
  • D. Create an AWS Glue crawler and custom classifier based upon the data feed formats and build a table definition to match. Perform an Amazon Athena query on file delivery to start an Amazon EMR ETL job to transform the entire record according to the processing and transformation requirements. Define the output format as JSON. Once complete, send the results to another S3 bucket for internal processing and scale down the EMR cluster.
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
God_Is_Love
Highly Voted 1 year, 7 months ago
Selected Answer: C
Extract Data from S3 + mask + Send to another S3 + Transform/Process + Load into S3 All these are ETL, ELT tasks which should ring Glue EMR is more focused on big data processing frameworks such as Hadoop and Spark, while Glue is more focused on ETL, More over 5000 records every 15 minutes is not soo big data..So I choose C
upvoted 21 times
tycho
1 year, 6 months ago
EMR and Glue are the same; Glue is managed cluster by AWS , EMR customer manages the clutster
upvoted 2 times
...
...
masetromain
Highly Voted 1 year, 9 months ago
Selected Answer: C
C is correct. It will process the data in batch mode using Glue ETL job which can handle large amount of data and can be scheduled to run periodically. This solution is also easily expandable for future feeds. A: It uses multiple Lambda functions, SQS queue and S3 temporary location which will increase operational overhead. B: Using Fargate may not be the most cost-effective solution and also it may not handle large amount of data. D: Athena and EMR both are powerful tools but they are more complex and can be more costly than Glue.
upvoted 7 times
...
career360guru
Most Recent 10 months, 1 week ago
Selected Answer: C
Option C
upvoted 1 times
...
totten
1 year ago
Selected Answer: C
Option C is the most suitable solution for the described scenario: 1) AWS Glue Crawler and Custom Classifier: Use AWS Glue to create a crawler and custom classifier to understand and catalogue the data feed formats. This step ensures that AWS Glue can work with the incoming data effectively. 2) AWS Glue ETL Job: Create an AWS Lambda function that triggers an AWS Glue ETL job when a new data file is delivered. This ETL job can perform the required transformation, including masking, field removal, and converting records to JSON format. AWS Glue is a suitable service for data preparation and transformation. 3) Output to S3 Bucket. This approach is scalable, easily expandable to handle additional feeds in the future, and leverages AWS Glue's capabilities for data transformation and processing. It also maintains a clear separation of tasks, making it a robust and efficient solution for the given requirements.
upvoted 3 times
...
dkcloudguru
1 year, 1 month ago
C is the good option EMR(Big data, Spark, Hadoop) is for near real-time data processing and it isn't a good fit in this case
upvoted 1 times
...
NikkyDicky
1 year, 3 months ago
Selected Answer: C
its a C
upvoted 1 times
...
SkyZeroZx
1 year, 4 months ago
Selected Answer: C
EMR is big data but not is need in this case then AWS Glue + Lambdas + S3 is good option C
upvoted 1 times
...
mfsec
1 year, 7 months ago
Selected Answer: C
C makes the most sense.
upvoted 2 times
...
Musk
1 year, 8 months ago
The question is at what point Athena and EMR are a better choice because it is a lot of data to store and process
upvoted 1 times
Sarutobi
1 year, 7 months ago
That, I agree. Honestly, I will use it from day one, regardless.
upvoted 1 times
...
...
zozza2023
1 year, 9 months ago
Selected Answer: C
C is correct.
upvoted 4 times
...
zhangyu20000
1 year, 9 months ago
C is correct
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago