exam questions

Exam AWS Certified Machine Learning - Specialty All Questions

View all questions & answers for the AWS Certified Machine Learning - Specialty exam

Exam AWS Certified Machine Learning - Specialty topic 1 question 176 discussion

A machine learning (ML) specialist needs to extract embedding vectors from a text series. The goal is to provide a ready-to-ingest feature space for a data scientist to develop downstream ML predictive models. The text consists of curated sentences in English. Many sentences use similar words but in different contexts. There are questions and answers among the sentences, and the embedding space must differentiate between them.
Which options can produce the required embedding vectors that capture word context and sequential QA information? (Choose two.)

  • A. Amazon SageMaker seq2seq algorithm
  • B. Amazon SageMaker BlazingText algorithm in Skip-gram mode
  • C. Amazon SageMaker Object2Vec algorithm
  • D. Amazon SageMaker BlazingText algorithm in continuous bag-of-words (CBOW) mode
  • E. Combination of the Amazon SageMaker BlazingText algorithm in Batch Skip-gram mode with a custom recurrent neural network (RNN)
Show Suggested Answer Hide Answer
Suggested Answer: AC 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
ovokpus
Highly Voted 2 years, 10 months ago
Selected Answer: AC
seq2seq and object2vec take care of more than just the words. Any response with blazingText is wrong because blazingText just uses a cbow (continuous bag of words), working only on individual words
upvoted 16 times
Jerry84
2 years, 3 months ago
“One of the well-known embedding techniques is Word2Vec, which provides embeddings for words.” “In addition to word embeddings, there are also use cases where we want to learn the embeddings of more general-purpose objects such as sentences, customers, and products. This is so we can build practical applications for information retrieval, product search, item matching, customer profiling based on similarity or as inputs for other supervised tasks. This is where Amazon SageMaker Object2Vec comes in.” https://aws.amazon.com/blogs/machine-learning/introduction-to-amazon-sagemaker-object2vec/
upvoted 1 times
...
robotgeek
1 year, 7 months ago
This is wrong, maybe the response (and the question) is outdated because BlazingText now supports three diferent techniques
upvoted 2 times
...
...
peterfish
Highly Voted 2 years, 9 months ago
Selected Answer: BD
It should be B an D. The objective is to create a latent space/word embedding that puts similar words closer to each other for other purposes. Thus, we should use Sagemaker Blazing Text in unsupervised mode (Word2Vec mode). cbow, skip-grams, and batch skip-grams are the 3 algorithms for this. However, since we do not need to do the later part of E, E is not correct. The ans should be B and D.
upvoted 7 times
DimLam
1 year, 6 months ago
yeah, my initial thought was the same. But both B and D embed words, not sentences.
upvoted 1 times
...
...
CloudGyan
Most Recent 3 months, 1 week ago
Selected Answer: CE
Using Amazon SageMaker Object2Vec (C) provides an end-to-end solution for learning embeddings with contextual and sequential relationships, while BlazingText in Skip-gram mode combined with a custom RNN (E) allows for greater flexibility in capturing sequence-level dependencies. Both approaches can produce high-quality embedding vectors that meet the requirements. Why not A : The seq2seq algorithm is designed for sequence-to-sequence tasks like translation or summarization. While it generates embeddings internally, it is not designed to provide a general-purpose feature space for downstream predictive models.
upvoted 1 times
...
GS_77
7 months, 2 weeks ago
Selected Answer: CE
Best choices
upvoted 1 times
...
AIWave
1 year, 2 months ago
Selected Answer: BC
To extract embedding vectors: Blazingtext Word2vec and Object2vec (B, C). Seq to seq: generate one sequence from another (A is out) Amazon SageMaker BlazingText algorithm in continuous bag-of-words (CBOW) mode does not capture word embeddings (D is out)
upvoted 2 times
...
kyuhuck
1 year, 2 months ago
Selected Answer: BE
To capture word context and sequential QA information, the embedding vectors need to consider both the order and the meaning of the words in the text. Option B, Amazon SageMaker BlazingText algorithm in Skip-gram mode, is a valid option because it can learn word embeddings that capture the semantic similarity and syntactic relations between words based on their co-occurrence in a window of words. Skip-gram mode can also handle rare words better than continuous bag-of-words (CBOW) mode1. Option E, combination of the Amazon SageMaker BlazingText algorithm in Batch Skip-gram mode with a custom recurrent neural network (RNN), is another valid option because it can leverage the advantages of Skip-gram mode and also use an RNN to model the sequential nature of the text. An RNN can capture the temporal dependencies and long-term dependencies between words, which are important for QA tasks2
upvoted 2 times
...
kyuhuck
1 year, 2 months ago
Selected Answer: CE
Considering the requirements, the two options that can produce the required embedding vectors that capture word context and sequential QA information are: C. Amazon SageMaker Object2Vec algorithm: Because it can learn to capture relationships in pairs of text, which could include the sequential nature of questions and answers. E. Combination of the Amazon SageMaker BlazingText algorithm in Batch Skip-gram mode with a custom recurrent neural network (RNN): This combination provides both context-aware word embeddings and the ability to capture sequential dependencies in text data.
upvoted 3 times
...
aakash_0086
1 year, 3 months ago
Selected Answer: CE
C because Object2Vec is a neural network-based algorithm that can learn embeddings for a wide range of data types and tasks. E because If you want to capture word context and sequential information, especially in the context of natural language processing (NLP), it is advisable to use models that are specifically designed for sequence modeling, such as recurrent neural networks (RNNs) or more advanced models like long short-term memory networks (LSTMs) or transformers.
upvoted 3 times
...
CloudHandsOn
1 year, 3 months ago
Selected Answer: AE
A. Amazon SageMaker seq2seq algorithm: Sequence-to-sequence (seq2seq) models are designed to convert sequences from one domain to sequences in another domain, often used in tasks like machine translation. They are capable of understanding the context and the sequence in which words appear, making them suitable for differentiating between questions and answers in a text. E. Combination of the Amazon SageMaker BlazingText algorithm in Batch Skip-gram mode with a custom recurrent neural network (RNN): This combination is promising. BlazingText in Skip-gram mode captures word context, and the recurrent neural network (RNN) is excellent for capturing sequential data, such as the flow in conversations or text. This combination should be effective at understanding both the context of individual words and the sequence of questions and answers.
upvoted 1 times
...
DimLam
1 year, 6 months ago
One problem of Object2Vec is it takes two objects as input during training and loss minimizes the difference between embeddings of these two objects. I don't think we have some labels to pass to Object2Vec We might think that we have a QA which we can pass as two objects. But in the question we want embeddings to distinguish between Q & A, but this Object2Vec minimizes the difference. So I wouldn't tell it is for sure C https://aws.amazon.com/blogs/machine-learning/introduction-to-amazon-sagemaker-object2vec/
upvoted 1 times
...
seifskl
1 year, 6 months ago
Selected Answer: AE
A . Seq2seq (sequence-to-sequence) models are designed to handle sequences. They are particularly well-suited for tasks like translating sentences from one language to another, but they can also be used for other tasks that involve sequences, such as converting questions to answers. Given that the embedding space must differentiate between questions and answers, a seq2seq model would be a good choice. E . BlazingText in Skip-gram mode can capture word context effectively. However, on its own, it might not capture the sequential information between questions and answers. By combining it with a custom RNN, the sequential nature of the sentences, especially in a QA setting, can be captured. RNNs are designed to work with sequences and can remember past information, making them suitable for this task.
upvoted 3 times
...
loict
1 year, 7 months ago
Selected Answer: BD
A. NO - seq2seq not for word embeddings B. YES - BlazingText in Skip-gram works and can capture Q&A C. NO - object2vec not for word embeddings D. YES - BlazingText in CBOW works and can capture Q&A E. NO - no need for RNN
upvoted 1 times
...
Shenannigan
1 year, 7 months ago
Selected Answer: BC
Answers should be B and C See the following for BlazingText with Skip-gram: https://arxiv.org/pdf/1604.04661.pdf (search skip-gram) Linked from this page https://docs.aws.amazon.com/sagemaker/latest/dg/blazingtext.html For Object2Vec see this page https://docs.aws.amazon.com/sagemaker/latest/dg/object2vec.html
upvoted 2 times
...
Mickey321
1 year, 8 months ago
Selected Answer: CE
C is confirmed but confused between A or E then lean to E
upvoted 2 times
...
Mickey321
1 year, 8 months ago
Selected Answer: CE
confusing
upvoted 1 times
...
kaike_reis
1 year, 8 months ago
Selected Answer: BC
Seq2Seq will not generate an embedding vector, so A it's wrong from my POV. I go with B - C
upvoted 3 times
...
Mickey321
1 year, 8 months ago
Selected Answer: BE
IMO E for sure then either B or C Combination of the Amazon SageMaker BlazingText algorithm in Batch Skip-gram mode with a custom recurrent neural network (RNN) is a more sophisticated approach that can be used to capture sequential QA information. This is because RNNs are able to learn long-term dependencies between words.
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago