Win IT Exam with Last Dumps 2024


Microsoft AI-102 Exam

Page 23/25
Viewing Questions 221 230 out of 241 Questions
92.00%

Question 221
You build a bot.
You create an Azure Bot resource.
You need to deploy the bot to Azure.
What else should you create?




Question 222
You are building a chatbot by using the Microsoft Bot Framework SDK. The bot will be used to accept food orders from customers and allow the customers to customize each food item.
You need to configure the bot to ask the user for additional input based on the type of item ordered. The solution must minimize development effort.
Which two types of dialogs should you use? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.




Question 223
HOTSPOT
-
You are building a chatbot by using the Microsoft Bot Framework SDK.
You use an object named UserProfile to store user profile information and an object named ConversationData to store information related to a conversation.
You create the following state accessors to store both objects in state.
AI-102_223Q_1.png related to the Microsoft AI-102 Exam
The state storage mechanism is set to Memory Storage.
For each of the following statements, select Yes if the statement is true. Otherwise select No.
NOTE: Each correct selection is worth one point.
AI-102_223Q_2.png related to the Microsoft AI-102 Exam
Image AI-102_223R.png related to the Microsoft AI-102 Exam




Question 224
DRAG DROP -
You are developing the smart e-commerce project.
You need to design the skillset to include the contents of PDFs in searches.
How should you complete the skillset design diagram? To answer, drag the appropriate services to the correct stages. Each service may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
AI-102_224Q.png related to the Microsoft AI-102 Exam
Image AI-102_224R.png related to the Microsoft AI-102 Exam



Box 1: Azure Blob storage -
At the start of the pipeline, you have unstructured text or non-text content (such as images, scanned documents, or JPEG files). Data must exist in an Azure data storage service that can be accessed by an indexer.
Box 2: Computer Vision API -
Scenario: Provide users with the ability to search insight gained from the images, manuals, and videos associated with the products.
The Computer Vision Read API is Azure's latest OCR technology (learn what's new) that extracts printed text (in several languages), handwritten text (English only), digits, and currency symbols from images and multi-page PDF documents.
Box 3: Translator API -
Scenario: Product descriptions, transcripts, and alt text must be available in English, Spanish, and Portuguese.
Box 4: Azure Files -
Scenario: Store all raw insight data that was generated, so the data can be processed later.
Incorrect Answers:
The custom vision API from Microsoft Azure learns to recognize specific content in imagery and becomes smarter with training and time.
Reference:
https://docs.microsoft.com/en-us/azure/search/cognitive-search-concept-intro
https://docs.microsoft.com/en-us/azure/cognitive-services/computer-vision/overview-ocr

Question 225
DRAG DROP -
You are planning the product creation project.
You need to recommend a process for analyzing videos.
Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order. (Choose four.)
Select and Place:
AI-102_225Q.png related to the Microsoft AI-102 Exam
Image AI-102_225R.png related to the Microsoft AI-102 Exam



Scenario: All videos must have transcripts that are associated to the video and included in product descriptions.
Product descriptions, transcripts, and alt text must be available in English, Spanish, and Portuguese.
Step 1: Upload the video to blob storage
Given a video or audio file, the file is first dropped into a Blob Storage. T
Step 2: Index the video by using the Video Indexer API.
When a video is indexed, Video Indexer produces the JSON content that contains details of the specified video insights. The insights include: transcripts, OCRs, faces, topics, blocks, etc.
Step 3: Extract the transcript from the Video Indexer API.
Step 4: Translate the transcript by using the Translator API.
Reference:
https://azure.microsoft.com/en-us/blog/get-video-insights-in-even-more-languages/
https://docs.microsoft.com/en-us/azure/media-services/video-indexer/video-indexer-output-json-v2


Question 226
You are developing the chatbot.
You create the following components:
- A QnA Maker resource
- A chatbot by using the Azure Bot Framework SDK
You need to integrate the components to meet the chatbot requirements.
Which property should you use?



Scenario: When the response confidence score is low, ensure that the chatbot can provide other response options to the customers.
When no good match is found by the ranker, the confidence score of 0.0 or "None" is returned and the default response is "No good match found in the KB". You can override this default response in the bot or application code calling the endpoint. Alternately, you can also set the override response in Azure and this changes the default for all knowledge bases deployed in a particular QnA Maker service.
Choosing Ranker type: By default, QnA Maker searches through questions and answers. If you want to search through questions only, to generate an answer, use the RankerType=QuestionOnly in the POST body of the GenerateAnswer request.
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/best-practices

Question 227
You are developing the chatbot.
You create the following components:
- A QnA Maker resource
- A chatbot by using the Azure Bot Framework SDK
You need to add an additional component to meet the technical requirements and the chatbot requirements.
What should you add?



Scenario: All planned projects must support English, French, and Portuguese.
If a bot uses multiple LUIS models and QnA Maker knowledge bases (knowledge bases), you can use the Dispatch tool to determine which LUIS model or QnA
Maker knowledge base best matches the user input. The dispatch tool does this by creating a single LUIS app to route user input to the correct model.
Reference:
https://docs.microsoft.com/en-us/azure/bot-service/bot-builder-tutorial-dispatch

Question 228
HOTSPOT -
You build a QnA Maker resource to meet the chatbot requirements.
Which RBAC role should you assign to each group? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
AI-102_228Q.png related to the Microsoft AI-102 Exam
Image AI-102_228R.png related to the Microsoft AI-102 Exam



Box 1: Cognitive Service User -
Ensure that the members of a group named Management-Accountants can approve the FAQs.
Approve=publish.
Cognitive Service User (read/write/publish): API permissions: All access to Cognitive Services resource except for ability to:
1. Add new members to roles.
2. Create new resources.
Box 2: Cognitive Services QnA Maker Editor
Ensure that the members of a group named Consultant-Accountants can create and amend the FAQs.
QnA Maker Editor: API permissions:
1. Create KB API
2. Update KB API
3. Replace KB API
4. Replace Alterations
5. "Train API" [in new service model v5]
Box 3: Cognitive Services QnA Maker Read
Ensure that the members of a group named the Agent-CustomerServices can browse the FAQs.
QnA Maker Read: API Permissions:
1. Download KB API
2. List KBs for user API
3. Get Knowledge base details
4. Download Alterations
Generate Answer -
Reference:
https://docs.microsoft.com/en-us/azure/cognitive-services/qnamaker/concepts/role-based-access-control

Question 229
DRAG DROP -
You are planning the product creation project.
You need to recommend a process for analyzing videos.
Which four actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:
AI-102_229Q.jpg related to the Microsoft AI-102 Exam
Image AI-102_229R.jpg related to the Microsoft AI-102 Exam



Scenario: All videos must have transcripts that are associated to the video and included in product descriptions.
Product descriptions, transcripts, and alt text must be available in English, Spanish, and Portuguese.
Step 1: Upload the video to blob storage
Given a video or audio file, the file is first dropped into a Blob Storage. T
Step 2: Index the video by using the Video Indexer API.
When a video is indexed, Video Indexer produces the JSON content that contains details of the specified video insights. The insights include: transcripts, OCRs, faces, topics, blocks, etc.
Step 3: Extract the transcript from the Video Indexer API.
Step 4: Translate the transcript by using the Translator API.
Reference:
https://azure.microsoft.com/en-us/blog/get-video-insights-in-even-more-languages/
https://docs.microsoft.com/en-us/azure/media-services/video-indexer/video-indexer-output-json-v2

Question 230
HOTSPOT -
You need to develop code to upload images for the product creation project. The solution must meet the accessibility requirements.
How should you complete the code? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
AI-102_230Q.png related to the Microsoft AI-102 Exam
Image AI-102_230R.png related to the Microsoft AI-102 Exam



Reference:
https://github.com/Azure-Samples/cognitive-services-dotnet-sdk-samples/blob/master/documentation-samples/quickstarts/ComputerVision/Program.cs





Premium Version