ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 158 - MLS-C01 discussion

Report
Export

A company uses a long short-term memory (LSTM) model to evaluate the risk factors of a particular energy sector. The model reviews multi-page text documents to analyze each sentence of the text and categorize it as either a potential risk or no risk. The model is not performing well, even though the Data Scientist has experimented with many different network structures and tuned the corresponding hyperparameters.

Which approach will provide the MAXIMUM performance boost?

A.
Initialize the words by term frequency-inverse document frequency (TF-IDF) vectors pretrained on a large collection of news articles related to the energy sector.
Answers
A.
Initialize the words by term frequency-inverse document frequency (TF-IDF) vectors pretrained on a large collection of news articles related to the energy sector.
B.
Use gated recurrent units (GRUs) instead of LSTM and run the training process until the validation loss stops decreasing.
Answers
B.
Use gated recurrent units (GRUs) instead of LSTM and run the training process until the validation loss stops decreasing.
C.
Reduce the learning rate and run the training process until the training loss stops decreasing.
Answers
C.
Reduce the learning rate and run the training process until the training loss stops decreasing.
D.
Initialize the words by word2vec embeddings pretrained on a large collection of news articles related to the energy sector.
Answers
D.
Initialize the words by word2vec embeddings pretrained on a large collection of news articles related to the energy sector.
Suggested answer: D

Explanation:

Initializing the words by word2vec embeddings pretrained on a large collection of news articles related to the energy sector will provide the maximum performance boost for the LSTM model. Word2vec is a technique that learns distributed representations of words based on their co-occurrence in a large corpus of text. These representations capture semantic and syntactic similarities between words, which can help the LSTM model better understand the meaning and context of the sentences in the text documents. Using word2vec embeddings that are pretrained on a relevant domain (energy sector) can further improve the performance by reducing the vocabulary mismatch and increasing the coverage of the words in the text documents.References:

AWS Machine Learning Specialty Exam Guide

AWS Machine Learning Training - Text Classification with TF-IDF, LSTM, BERT: a comparison of performance

AWS Machine Learning Training - Machine Learning - Exam Preparation Path

asked 16/09/2024
Oleksii Ivanov
45 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first