ExamGecko
Question list
Search
Search

List of questions

Search

Question 34 - H13-311_V3.5 discussion

Report
Export

Sigmoid, tanh, and softsign activation functions cannot avoid vanishing gradient problems when the network is deep.

A.
TRUE
Answers
A.
TRUE
B.
FALSE
Answers
B.
FALSE
Suggested answer: A

Explanation:

Activation functions like Sigmoid, tanh, and softsign suffer from the vanishing gradient problem when used in deep networks. This happens because, in these functions, gradients become very small as the input moves away from the origin (either positively or negatively). As a result, the weights of the earlier layers in the network receive very small updates, hindering the learning process in deep networks. This is one reason why activation functions like ReLU, which avoid this issue, are often preferred in deep learning.

asked 26/09/2024
Stergios Gaidatzis
38 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first