Huawei H13-311_V3.5 Practice Test - Questions Answers, Page 3

List of questions
Question 21

An algorithm of unsupervised learning classifies samples in a dataset into several categories. Samples belonging to the same category have high similarity.
In unsupervised learning, the goal is to find hidden patterns or intrinsic structures in input data without labeled outcomes. One common unsupervised learning task is clustering, where an algorithm groups the dataset into several categories or clusters. Samples within the same cluster have high similarity based on certain features, while samples in different clusters have low similarity. Examples of clustering algorithms include k-means and hierarchical clustering.
Question 22

Which of the following statements is false about the debugging and application of a regression model?
Logistic regression is not a solution for underfitting in regression models, as it is used primarily for classification problems rather than regression tasks. If underfitting occurs, it means that the model is too simple to capture the underlying patterns in the data. Solutions include using a more complex regression model like polynomial regression or increasing the number of features in the dataset.
Other options like adding a regularization term for overfitting (Lasso or Ridge) and using data cleansing and feature engineering are correct methods for improving model performance.
Question 23

In machine learning, which of the following inputs is required for model training and prediction?
In machine learning, historical data is crucial for model training and prediction. The model learns from this data, identifying patterns and relationships between features and target variables. While the training algorithm is necessary for defining how the model learns, the input required for the model is historical data, as it serves as the foundation for training the model to make future predictions.
Neural networks and training algorithms are parts of the model development process, but they are not the actual input for model training.
Question 24

Which of the following statements about datasets are true?
In machine learning:
The testing set is a dataset used after training to evaluate the model's performance and generalization ability. Each sample in this set is called a test sample.
A dataset generally has multiple dimensions, with each dimension representing a feature or attribute of the data.
A typical machine learning process divides the data into a training set (to train the model), a validation set (to tune hyperparameters and avoid overfitting), and a test set (to evaluate the model's final performance).
The statement that the validation set and test set are the same is false because they serve different purposes: validation is for hyperparameter tuning, while testing is for final model evaluation.
Question 25

In a hyperparameter-based search, the hyperparameters of a model are searched based on the data on and the model's performance metrics.
In machine learning, hyperparameters are the parameters that govern the learning process and are not learned from the data. Hyperparameter optimization or hyperparameter tuning is a critical part of improving a model's performance. The goal of a hyperparameter-based search is to find the set of hyperparameters that maximizes the model's performance on a given dataset.
There are different techniques for hyperparameter tuning, such as grid search, random search, and more advanced methods like Bayesian optimization. The performance of the model is assessed based on evaluation metrics (like accuracy, precision, recall, etc.), and the hyperparameters are adjusted accordingly to achieve the best performance.
In Huawei's HCIA AI curriculum, hyperparameter optimization is discussed in relation to both traditional machine learning models and deep learning frameworks. The course emphasizes the importance of selecting appropriate hyperparameters and demonstrates how frameworks such as TensorFlow and Huawei's ModelArts platform can facilitate hyperparameter searches to optimize models efficiently.
HCIA AI
AI Overview and Machine Learning Overview: Emphasize the importance of hyperparameters in model training.
Deep Learning Overview: Highlights the role of hyperparameter tuning in neural network architectures, including tuning learning rates, batch sizes, and other key parameters.
AI Development Frameworks: Discusses the use of hyperparameter search tools in platforms like TensorFlow and Huawei ModelArts.
Question 26

Fill in blanks
The general process of building a project using machine learning involves the following steps: split data, _________________ the model, deploy the model the model, and fine-tune the model.
Question 27

When feature engineering is complete, which of the following is not a step in the decision tree building process?
When building a decision tree, the steps generally involve:
Decision tree generation: This is the process where the model iteratively splits the data based on feature values to form branches.
Pruning: This step occurs post-generation, where unnecessary branches are removed to reduce overfitting and enhance generalization.
Feature selection: This is part of decision tree construction, where relevant features are selected at each node to determine how the tree branches.
Data cleansing, on the other hand, is a preprocessing step carried out before any model training begins. It involves handling missing or erroneous data to improve the quality of the dataset but is not part of the decision tree building process itself.
HCIA AI
Machine Learning Overview: Includes a discussion on decision tree algorithms and the process of building decision trees.
AI Development Framework: Highlights the steps for building machine learning models, separating data preprocessing (e.g., data cleansing) from model building steps.
Question 28

Which of the following statements are true about decision trees?
A . TRUE. The common decision tree algorithms include ID3, C4.5, and CART. These are the most widely used algorithms for decision tree generation.
B . FALSE. Purity in decision trees can be measured using multiple metrics, such as information gain, Gini index, and others, not just information entropy.
C . TRUE. Building a decision tree involves selecting the best features and determining their order in the tree structure to split the data effectively.
D . TRUE. One key step in decision tree generation is evaluating the purity of different splits (e.g., how well the split segregates the target variable) by comparing metrics like information gain or Gini index.
HCIA AI
Machine Learning Overview: Covers decision tree algorithms and their use cases.
Deep Learning Overview: While this focuses on neural networks, it touches on how decision-making algorithms are used in structured data models.
Question 29

The training error decreases as the model complexity increases.
As the model complexity increases (for example, by adding more layers to a neural network or increasing the depth of a decision tree), the training error tends to decrease. This is because more complex models are able to fit the training data better, possibly even capturing noise. However, increasing complexity often leads to overfitting, where the model performs well on the training data but poorly on unseen test data.
The relationship between model complexity and performance is covered extensively in Huawei HCIA AI's discussion of overfitting and underfitting and how model generalization is affected by increasing model complexity.
HCIA AI
Machine Learning Overview: Explains model complexity and its effect on training and testing error curves.
Deep Learning Overview: Discusses the balance between model capacity, overfitting, and underfitting in deep learning architectures.
Question 30

Which of the following is the activation function used in the hidden layers of the standard recurrent neural network (RNN) structure?
In standard Recurrent Neural Networks (RNNs), the Tanh activation function is commonly used in the hidden layers. The Tanh function squashes input values to a range between -1 and 1, allowing the network to learn complex patterns over time by transforming the input data into non-linear patterns.
While other activation functions like Sigmoid can be used, Tanh is preferred in many RNNs for its wider range. ReLU is generally used in feed-forward networks, and Softmax is often applied in the output layer for classification problems.
HCIA AI
Deep Learning Overview: Describes the architecture of RNNs, highlighting the use of Tanh as the standard activation function.
AI Development Framework: Discusses the various activation functions used across different neural network architectures.
Question