Microsoft AI-900 Practice Test - Questions Answers, Page 3

List of questions
Question 21

You need to predict the income range of a given customer by using the following dataset.
Which two fields should you use as features? Each correct answer presents a complete solution.
NOTE: Each correct selection is worth one point.
First Name, Last Name, Age and Education Level are features. Income range is a label (what you want to predict). First Name and Last Name are irrelevant in that they have no bearing on income. Age and Education level are the features you should use.
Question 22

You need to develop a mobile app for employees to scan and store their expenses while travelling.
Which type of computer vision should you use?
Azure's Computer Vision API includes Optical Character Recognition (OCR) capabilities that extract printed or handwritten text from images. You can extract text from images, such as photos of license plates or containers with serial numbers, as well as from documents - invoices, bills, financial reports, articles, and more.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/concept-recognizing-text
Question 23

You need to determine the location of cars in an image so that you can estimate the distance between the cars. Which type of computer vision should you use?
Object detection is similar to tagging, but the API returns the bounding box coordinates (in pixels) for each object found. For example, if an image contains a dog, cat and person, the Detect operation will list those objects together with their coordinates in the image. You can use this functionality to process the relationships between the objects in an image. It also lets you determine whether there are multiple instances of the same tag in an image. The Detect API applies tags based on the objects or living things identified in the image. There is currently no formal relationship between the tagging taxonomy and the object detection taxonomy. At a conceptual level, the Detect API only finds objects and living things, while the Tag API can also include contextual terms like "indoor", which can't be localized with bounding boxes.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/concept-object-detection
Question 24

You send an image to a Computer Vision API and receive back the annotated image shown in the exhibit.
Which type of computer vision was used?
Object detection is similar to tagging, but the API returns the bounding box coordinates (in pixels) for each object found. For example, if an image contains a dog, cat and person, the Detect operation will list those objects together with their coordinates in the image. You can use this functionality to process the relationships between the objects in an image. It also lets you determine whether there are multiple instances of the same tag in an image. The Detect API applies tags based on the objects or living things identified in the image. There is currently no formal relationship between the tagging taxonomy and the object detection taxonomy. At a conceptual level, the Detect API only finds objects and living things, while the Tag API can also include contextual terms like "indoor", which can't be localized with bounding boxes.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/concept-object-detection
Question 25

What are two tasks that can be performed by using the Computer Vision service? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
B: Azure's Computer Vision service provides developers with access to advanced algorithms that process images and return information based on the visual features you're interested in. For example, Computer Vision can determine whether an image contains adult content, find specific brands or objects, or find human faces.
C: Computer Vision includes Optical Character Recognition (OCR) capabilities. You can use the new Read API to extract printed and handwritten text from images and documents.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/home
Question 26

What is a use case for classification?
Question 27

What are two tasks that can be performed by using computer vision? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
B: Identify commercial brands in images or videos from a database of thousands of global logos. You can use this feature, for example, to discover which brands are most popular on social media or most prevalent in media product placement.
C: Analyze color usage within an image. Computer Vision can determine whether an image is black & white or color and, for color images, identify the dominant and accent colors.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/overview
Question 28

Your company wants to build a recycling machine for bottles. The recycling machine must automatically identify bottles of the correct shape and reject all other items. Which type of AI workload should the company use?
Azure's Computer Vision service gives you access to advanced algorithms that process images and return information based on the visual features you're interested in. For example, Computer Vision can determine whether an image contains adult content, find specific brands or objects, or find human faces.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/overview
Question 29

In which two scenarios can you use the Form Recognizer service? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.
Reference:
https://azure.microsoft.com/en-gb/services/cognitive-services/form-recognizer/#features
Question 30

Your website has a chatbot to assist customers.
You need to detect when a customer is upset based on what the customer types in the chatbot.
Which type of AI workload should you use?
Natural language processing (NLP) is used for tasks such as sentiment analysis, topic detection, language detection, key phrase extraction, and document categorization.
Sentiment Analysis is the process of determining whether a piece of writing is positive, negative or neutral.
Reference:
https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/natural-language-processing
Question