We need to first vectorize the user question, these vectors will then be used to search the database for relevant documents which will create an augmented prompt that the response-generating LLM will use to provide an answer
In a typical RAG-enabled chatbot, the process usually follows these steps:
Embedding Model: Converts the user’s question into a vector representation.
Vector Search: Finds relevant information based on the vector representation.
Context-Augmented Prompt: Combines the retrieved information with the original question to create a prompt.
Response-Generating LLM: Generates the final response based on the context-augmented prompt.
So, the correct sequence is:
A. 1.embedding model, 2.vector search, 3.context-augmented prompt, 4.response-generating LLM
Option C suggests starting with the response-generating LLM, which doesn’t align with the typical RAG workflow. The LLM needs the context-augmented prompt to generate a relevant response, which is why it comes last in the sequence.
upvoted 1 times
...
Log in to ExamTopics
Sign in:
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.
Upvoting a comment with a selected answer will also increase the vote count towards that answer by one.
So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.
4af18fc
Highly Voted 1 month, 2 weeks agotrendy01
Most Recent 4 weeks, 1 day agoHemaKG
1 month ago