ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 197 - MLS-C01 discussion

Report
Export

A data scientist has developed a machine learning translation model for English to Japanese by using Amazon SageMaker's built-in seq2seq algorithm with 500,000 aligned sentence pairs. While testing with sample sentences, the data scientist finds that the translation quality is reasonable for an example as short as five words. However, the quality becomes unacceptable if the sentence is 100 words long.

Which action will resolve the problem?

A.
Change preprocessing to use n-grams.
Answers
A.
Change preprocessing to use n-grams.
B.
Add more nodes to the recurrent neural network (RNN) than the largest sentence's word count.
Answers
B.
Add more nodes to the recurrent neural network (RNN) than the largest sentence's word count.
C.
Adjust hyperparameters related to the attention mechanism.
Answers
C.
Adjust hyperparameters related to the attention mechanism.
D.
Choose a different weight initialization type.
Answers
D.
Choose a different weight initialization type.
Suggested answer: C

Explanation:

The data scientist should adjust hyperparameters related to the attention mechanism to resolve the problem. The attention mechanism is a technique that allows the decoder to focus on different parts of the input sequence when generating the output sequence. It helps the model cope with long input sequences and improve the translation quality. The Amazon SageMaker seq2seq algorithm supports different types of attention mechanisms, such as dot, general, concat, and mlp. The data scientist can use the hyperparameter attention_type to choose the type of attention mechanism. The data scientist can also use the hyperparameter attention_coverage_type to enable coverage, which is a mechanism that penalizes the model for attending to the same input positions repeatedly. By adjusting these hyperparameters, the data scientist can fine-tune the attention mechanism and reduce the number of false negative predictions by the model.

References:

Sequence-to-Sequence Algorithm - Amazon SageMaker

Attention Mechanism - Sockeye Documentation

asked 16/09/2024
Lance Gentle
45 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first