exam questions

Exam AWS Certified Data Analytics - Specialty All Questions

View all questions & answers for the AWS Certified Data Analytics - Specialty exam

Exam AWS Certified Data Analytics - Specialty topic 1 question 80 discussion

A company has an application that ingests streaming data. The company needs to analyze this stream over a 5-minute timeframe to evaluate the stream for anomalies with Random Cut Forest (RCF) and summarize the current count of status codes. The source and summarized data should be persisted for future use.
Which approach would enable the desired outcome while keeping data persistence costs low?

  • A. Ingest the data stream with Amazon Kinesis Data Streams. Have an AWS Lambda consumer evaluate the stream, collect the number status codes, and evaluate the data against a previously trained RCF model. Persist the source and results as a time series to Amazon DynamoDB.
  • B. Ingest the data stream with Amazon Kinesis Data Streams. Have a Kinesis Data Analytics application evaluate the stream over a 5-minute window using the RCF function and summarize the count of status codes. Persist the source and results to Amazon S3 through output delivery to Kinesis Data Firehouse.
  • C. Ingest the data stream with Amazon Kinesis Data Firehose with a delivery frequency of 1 minute or 1 MB in Amazon S3. Ensure Amazon S3 triggers an event to invoke an AWS Lambda consumer that evaluates the batch data, collects the number status codes, and evaluates the data against a previously trained RCF model. Persist the source and results as a time series to Amazon DynamoDB.
  • D. Ingest the data stream with Amazon Kinesis Data Firehose with a delivery frequency of 5 minutes or 1 MB into Amazon S3. Have a Kinesis Data Analytics application evaluate the stream over a 1-minute window using the RCF function and summarize the count of status codes. Persist the results to Amazon S3 through a Kinesis Data Analytics output to an AWS Lambda integration.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Donell
Highly Voted 3 years, 5 months ago
Answer is B. First of all the question mentions about analyzing the stream over a 5-minute timeframe. Only option B satisfies the above condition. Also KDA uses RCF. Q: How can I perform real-time anomaly detection in Kinesis Data Analytics? Kinesis Data Analytics includes pre-built SQL functions for several advanced analytics including one for anomaly detection. You can simply make a call to this function from your SQL code for detecting anomalies in real-time. Kinesis Data Analytics uses the Random Cut Forest algorithm to implement anomaly detection. For more information on Random Cut Forests, see the Streaming Data Anomaly Detection whitepaper.
upvoted 20 times
...
VikG12
Highly Voted 3 years, 6 months ago
Looks like 'B' it is.
upvoted 6 times
...
weasd555
Most Recent 1 year, 6 months ago
Answer is . B Ingest the data stream with Amazon Kinesis Data Streams. Have a Kinesis Data Analytics application evaluate the stream over a 5-minute window using the RCF function and summarize the count of status codes. Persist the source and results to Amazon S3 through output delivery to Kinesis Data Firehouse
upvoted 1 times
...
pk349
1 year, 11 months ago
B: I passed the test
upvoted 1 times
...
OLANGA
2 years, 4 months ago
Why further questions are not available?
upvoted 1 times
GoForTheWin
2 years, 2 months ago
Because you don't pay for further pages.
upvoted 1 times
...
...
cloudlearnerhere
2 years, 5 months ago
Selected Answer: B
Correct answer is B as Kinesis Data Streams can be used to capture the data with Kinesis Data Analytics to perform RCF over a 5 minutes window pushing the source and results data into Kinesis Data Firehose for S3 persistence. https://aws.amazon.com/blogs/big-data/perform-near-real-time-analytics-on-streaming-data-with-amazon-kinesis-and-amazon-elasticsearch-service/ Option A is wrong as it would require more effort and the cost would be higher if storing the data in DynamoDB. Option C is wrong as Kinesis Data Firehose with a delivery frequency of 1 minute or 1 MB would not be able to meet the requirement accurately. Also, the cost would be higher if storing the data in DynamoDB. Option D is wrong as having the Kinesis Data Analytics application evaluate the stream over a 1-minute window would not meet the 5 minutes requirement.
upvoted 5 times
...
t47
2 years, 5 months ago
answer is B
upvoted 2 times
...
awsdatacert
2 years, 6 months ago
Answer is B
upvoted 1 times
...
rocky48
2 years, 8 months ago
Selected Answer: B
Answer is B
upvoted 1 times
...
Heer
3 years, 5 months ago
ANSWER:OPTION B EXPLAINATION: Since it's a real time streaming data ,so filter out option C,D as KFH is near real time .Now out of option A and B ,S3 has a low data persistence cost .
upvoted 3 times
Donell
3 years, 5 months ago
Answer is B. However no where it is mentioned about real-time in question. Streaming Data can also be near real-time.Ruling out KDF due to the above reason is incorrect.
upvoted 6 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago