exam questions

Exam AWS Certified Solutions Architect - Professional All Questions

View all questions & answers for the AWS Certified Solutions Architect - Professional exam

Exam AWS Certified Solutions Architect - Professional topic 1 question 699 discussion

A life sciences company is using a combination of open source tools to manage data analysis workflows and Docker containers running on servers in its on- premises data center to process genomics data. Sequencing data is generated and stored on a local storage area network (SAN), and then the data is processed.
The research and development teams are running into capacity issues and have decided to re-architect their genomics analysis platform on AWS to scale based on workload demands and reduce the turnaround time from weeks to days.
The company has a high-speed AWS Direct Connect connection. Sequencers will generate around 200 GB of data for each genome, and individual jobs can take several hours to process the data with ideal compute capacity. The end result will be stored in Amazon S3. The company is expecting 10-15 job requests each day.
Which solution meets these requirements?

  • A. Use regularly scheduled AWS Snowball Edge devices to transfer the sequencing data into AWS. When AWS receives the Snowball Edge device and the data is loaded into Amazon S3, use S3 events to trigger an AWS Lambda function to process the data.
  • B. Use AWS Data Pipeline to transfer the sequencing data to Amazon S3. Use S3 events to trigger an Amazon EC2 Auto Scaling group to launch custom-AMI EC2 instances running the Docker containers to process the data.
  • C. Use AWS DataSync to transfer the sequencing data to Amazon S3. Use S3 events to trigger an AWS Lambda function that starts an AWS Step Functions workflow. Store the Docker images in Amazon Elastic Container Registry (Amazon ECR) and trigger AWS Batch to run the container and process the sequencing data.
  • D. Use an AWS Storage Gateway file gateway to transfer the sequencing data to Amazon S3. Use S3 events to trigger an AWS Batch job that executes on Amazon EC2 instances running the Docker containers to process the data.
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
CarisB
Highly Voted 3 years, 6 months ago
Agree on C. For instance: https://docs.aws.amazon.com/whitepapers/latest/genomics-data-transfer-analytics-and-machine-learning/transferring-genomics-data-to-the-cloud-and-establishing-data-access-patterns-using-aws-datasync-and-aws-storage-gateway-for-files.html => Use AWS DataSync to transfer data to Amazon S3
upvoted 22 times
ExtHo
3 years, 6 months ago
Very good reference provided clears doubts :)
upvoted 1 times
...
...
sek12324
Highly Voted 3 years, 7 months ago
C for me, as docker images are used - they need ECR
upvoted 15 times
...
Sin_Dan
Most Recent 6 months, 1 week ago
Selected Answer: C
Although option C is right, adding Lambda function to trigger the step function workflow is unnecessary. The step function can directly be triggered by the S3 event trigger.
upvoted 1 times
...
sjpd10
2 years, 6 months ago
The question has some key words that are referred in the AWS link - third-party tools, open-source tools, which is in the question. It also says, if the data is local storage (SAN), use 'Storage GW. Is the answer still A. I vote D.
upvoted 1 times
...
JohnPi
2 years, 6 months ago
the problem that I see with option C AWS DataSync is related to the fact that input is a storage area network (SAN) and according to the docs AWS DataSync can connect to NAS as a NFS/SMB file share
upvoted 2 times
JohnPi
2 years, 6 months ago
this is the statement: "Sequencing data is generated and stored on a local storage area network (SAN), and then the data is processed." so SAN not NAS
upvoted 1 times
...
...
asfsdfsdf
2 years, 9 months ago
Selected Answer: C
Answer is C - docker images with ECR / DataSync to move the data with DX / Batch to run compute process with containers - this done with job Type: "container" and an image from ECR
upvoted 1 times
...
johnnsmith
3 years, 2 months ago
D is the correct answer. C is incorrect. S3 event can trigger Batch job.
upvoted 2 times
asfsdfsdf
2 years, 9 months ago
This is incorrect, S3 event can trigger only the below services: Amazon Simple Notification Service (Amazon SNS) topics Amazon Simple Queue Service (Amazon SQS) queues AWS Lambda function Answer is C - docker images - ECR / DataSync to move the data with DX
upvoted 4 times
...
...
jj22222
3 years, 2 months ago
Selected Answer: C
CCCCCCCCCCCCCCCCCC
upvoted 2 times
...
pititcu667
3 years, 3 months ago
Selected Answer: C
c because i belive it's better to use ecs than lambda.
upvoted 2 times
...
cldy
3 years, 3 months ago
C is correct.
upvoted 1 times
...
mm84
3 years, 4 months ago
Agree on C.
upvoted 1 times
...
AzureDP900
3 years, 4 months ago
C is right
upvoted 1 times
...
andylogan
3 years, 5 months ago
It's C refer to compare DataSync usage to others https://aws.amazon.com/datasync/faqs/#When_to_choose_AWS_DataSync
upvoted 3 times
...
tgv
3 years, 5 months ago
CCC ---
upvoted 1 times
...
blackgamer
3 years, 6 months ago
C is the answer
upvoted 1 times
...
WhyIronMan
3 years, 6 months ago
I'll go with C
upvoted 2 times
...
WhyIronMan
3 years, 6 months ago
I'll go with C
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago