exam questions

Exam AWS Certified AI Practitioner AIF-C01 All Questions

View all questions & answers for the AWS Certified AI Practitioner AIF-C01 exam

Exam AWS Certified AI Practitioner AIF-C01 topic 1 question 125 discussion

HOTSPOT
-

A company has developed a large language model (LLM) and wants to make the LLM available to multiple internal teams. The company needs to select the appropriate inference mode for each team.

Select the correct inference mode from the following list for each use case. Each inference mode should be selected one or more times.

Show Suggested Answer Hide Answer
Suggested Answer:

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
4729e6c
12 hours, 42 minutes ago
In Scenario 1, the company's chatbot requires immediate predictions to accurately interpret user intent during interactions. Implementing real-time inference ensures that the model processes each user input instantaneously, providing timely and relevant responses essential for effective communication. For Scenario 2, the data processing job involves handling large volumes of text data during specific periods, such as weekends. Utilizing batch transform allows the system to process this data in bulk, optimizing resource utilization and efficiency without the need for immediate results. In Scenario 3, the engineering team aims to develop an API capable of swiftly processing small text inputs and delivering prompt predictions. Adopting real-time inference enables the API to handle each request as it arrives, ensuring low-latency responses critical for user-facing applications.
upvoted 1 times
...
Jessiii
1 day, 15 hours ago
real time batch real time chatbot (real-time, low-latency predictions): Real-time inference Data processing job (large datasets on weekends): Batch transform
upvoted 1 times
...
LonghornFan
2 days, 18 hours ago
Scenario 1: The company’s chatbot needs predictions from the LLM to understand users’ intent with minimal latency. Answer: Real-time inference Scenario 2: A data processing job needs to query the LLM to process gigabytes of text files on weekends. Answer: Batch transform Scenario 3: The company’s engineering team needs to create an API that can process small pieces of text content and provide low-latency predictions. Answer: Real-time inference
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago