Summer Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

AIF-C01 Exam Dumps - AWS Certified AI Practitioner Exam(AI1-C01)

Question # 4

An AI practitioner is building a model to generate images of humans in various professions. The AI practitioner discovered that the input data is biased and that specific attributes affect the image generation and create bias in the model.

Which technique will solve the problem?

A.

Data augmentation for imbalanced classes

B.

Model monitoring for class distribution

C.

Retrieval Augmented Generation (RAG)

D.

Watermark detection for images

Full Access
Question # 5

An AI company periodically evaluates its systems and processes with the help of independent software vendors (ISVs). The company needs to receive email message notifications when an ISV's compliance reports become available.

Which AWS service can the company use to meet this requirement?

A.

AWS Audit Manager

B.

AWS Artifact

C.

AWS Trusted Advisor

D.

AWS Data Exchange

Full Access
Question # 6

A company needs to choose a model from Amazon Bedrock to use internally. The company must identify a model that generates responses in a style that the company's employees prefer.

What should the company do to meet these requirements?

A.

Evaluate the models by using built-in prompt datasets.

B.

Evaluate the models by using a human workforce and custom prompt datasets.

C.

Use public model leaderboards to identify the model.

D.

Use the model InvocationLatency runtime metrics in Amazon CloudWatch when trying models.

Full Access
Question # 7

A company is building a chatbot to improve user experience. The company is using a large language model (LLM) from Amazon Bedrock for intent detection. The company wants to use few-shot learning to improve intent detection accuracy.

Which additional data does the company need to meet these requirements?

A.

Pairs of chatbot responses and correct user intents

B.

Pairs of user messages and correct chatbot responses

C.

Pairs of user messages and correct user intents

D.

Pairs of user intents and correct chatbot responses

Full Access
Question # 8

A company uses Amazon SageMaker for its ML pipeline in a production environment. The company has large input data sizes up to 1 GB and processing times up to 1 hour. The company needs near real-time latency.

Which SageMaker inference option meets these requirements?

A.

Real-time inference

B.

Serverless inference

C.

Asynchronous inference

D.

Batch transform

Full Access
Question # 9

A company is using an Amazon Bedrock base model to summarize documents for an internal use case. The company trained a custom model to improve the summarization quality.

Which action must the company take to use the custom model through Amazon Bedrock?

A.

Purchase Provisioned Throughput for the custom model.

B.

Deploy the custom model in an Amazon SageMaker endpoint for real-time inference.

C.

Register the model with the Amazon SageMaker Model Registry.

D.

Grant access to the custom model in Amazon Bedrock.

Full Access
Question # 10

An AI practitioner trained a custom model on Amazon Bedrock by using a training dataset that contains confidential data. The AI practitioner wants to ensure that the custom model does not generate inference responses based on confidential data.

How should the AI practitioner prevent responses based on confidential data?

A.

Delete the custom model. Remove the confidential data from the training dataset. Retrain the custom model.

B.

Mask the confidential data in the inference responses by using dynamic data masking.

C.

Encrypt the confidential data in the inference responses by using Amazon SageMaker.

D.

Encrypt the confidential data in the custom model by using AWS Key Management Service (AWS KMS).

Full Access
Question # 11

A company wants to build an ML model by using Amazon SageMaker. The company needs to share and manage variables for model development across multiple teams.

Which SageMaker feature meets these requirements?

A.

Amazon SageMaker Feature Store

B.

Amazon SageMaker Data Wrangler

C.

Amazon SageMaker Clarify

D.

Amazon SageMaker Model Cards

Full Access
Question # 12

A company wants to make a chatbot to help customers. The chatbot will help solve technical problems without human intervention. The company chose a foundation model (FM) for the chatbot. The chatbot needs to produce responses that adhere to company tone.

Which solution meets these requirements?

A.

Set a low limit on the number of tokens the FM can produce.

B.

Use batch inferencing to process detailed responses.

C.

Experiment and refine the prompt until the FM produces the desired responses.

D.

Define a higher number for the temperature parameter.

Full Access
Question # 13

Which AWS service or feature can help an AI development team quickly deploy and consume a foundation model (FM) within the team's VPC?

A.

Amazon Personalize

B.

Amazon SageMaker JumpStart

C.

PartyRock, an Amazon Bedrock Playground

D.

Amazon SageMaker endpoints

Full Access
Question # 14

A company uses a foundation model (FM) from Amazon Bedrock for an AI search tool. The company wants to fine-tune the model to be more accurate by using the company's data.

Which strategy will successfully fine-tune the model?

A.

Provide labeled data with the prompt field and the completion field.

B.

Prepare the training dataset by creating a .txt file that contains multiple lines in .csv format.

C.

Purchase Provisioned Throughput for Amazon Bedrock.

D.

Train the model on journals and textbooks.

Full Access
Question # 15

A company has installed a security camera. The company uses an ML model to evaluate the security camera footage for potential thefts. The company has discovered that the model disproportionately flags people who are members of a specific ethnic group.

Which type of bias is affecting the model output?

A.

Measurement bias

B.

Sampling bias

C.

Observer bias

D.

Confirmation bias

Full Access
Question # 16

A company wants to deploy a conversational chatbot to answer customer questions. The chatbot is based on a fine-tuned Amazon SageMaker JumpStart model. The application must comply with multiple regulatory frameworks.

Which capabilities can the company show compliance for? (Select TWO.)

A.

Auto scaling inference endpoints

B.

Threat detection

C.

Data protection

D.

Cost optimization

E.

Loosely coupled microservices

Full Access
Question # 17

A company is using a pre-trained large language model (LLM) to build a chatbot for product recommendations. The company needs the LLM outputs to be short and written in a specific language.

Which solution will align the LLM response quality with the company's expectations?

A.

Adjust the prompt.

B.

Choose an LLM of a different size.

C.

Increase the temperature.

D.

Increase the Top K value.

Full Access
Question # 18

A company wants to use a large language model (LLM) to develop a conversational agent. The company needs to prevent the LLM from being manipulated with common prompt engineering techniques to perform undesirable actions or expose sensitive information.

Which action will reduce these risks?

A.

Create a prompt template that teaches the LLM to detect attack patterns.

B.

Increase the temperature parameter on invocation requests to the LLM.

C.

Avoid using LLMs that are not listed in Amazon SageMaker.

D.

Decrease the number of input tokens on invocations of the LLM.

Full Access
Question # 19

A company wants to display the total sales for its top-selling products across various retail locations in the past 12 months.

Which AWS solution should the company use to automate the generation of graphs?

A.

Amazon Q in Amazon EC2

B.

Amazon Q Developer

C.

Amazon Q in Amazon QuickSight

D.

Amazon Q in AWS Chatbot

Full Access
Question # 20

An AI practitioner is using an Amazon Bedrock base model to summarize session chats from the customer service department. The AI practitioner wants to store invocation logs to monitor model input and output data.

Which strategy should the AI practitioner use?

A.

Configure AWS CloudTrail as the logs destination for the model.

B.

Enable invocation logging in Amazon Bedrock.

C.

Configure AWS Audit Manager as the logs destination for the model.

D.

Configure model invocation logging in Amazon EventBridge.

Full Access
Question # 21

How can companies use large language models (LLMs) securely on Amazon Bedrock?

A.

Design clear and specific prompts. Configure AWS Identity and Access Management (IAM) roles and policies by using least privilege access.

B.

Enable AWS Audit Manager for automatic model evaluation jobs.

C.

Enable Amazon Bedrock automatic model evaluation jobs.

D.

Use Amazon CloudWatch Logs to make models explainable and to monitor for bias.

Full Access
Question # 22

A company wants to use a pre-trained generative AI model to generate content for its marketing campaigns. The company needs to ensure that the generated content aligns with the company's brand voice and messaging requirements.

Which solution meets these requirements?

A.

Optimize the model's architecture and hyperparameters to improve the model's overall performance.

B.

Increase the model's complexity by adding more layers to the model's architecture.

C.

Create effective prompts that provide clear instructions and context to guide the model's generation.

D.

Select a large, diverse dataset to pre-train a new generative model.

Full Access
Question # 23

A medical company is customizing a foundation model (FM) for diagnostic purposes. The company needs the model to be transparent and explainable to meet regulatory requirements.

Which solution will meet these requirements?

A.

Configure the security and compliance by using Amazon Inspector.

B.

Generate simple metrics, reports, and examples by using Amazon SageMaker Clarify.

C.

Encrypt and secure training data by using Amazon Macie.

D.

Gather more data. Use Amazon Rekognition to add custom labels to the data.

Full Access
Question # 24

A company wants to assess the costs that are associated with using a large language model (LLM) to generate inferences. The company wants to use Amazon Bedrock to build generative AI applications.

Which factor will drive the inference costs?

A.

Number of tokens consumed

B.

Temperature value

C.

Amount of data used to train the LLM

D.

Total training time

Full Access
Question # 25

A company wants to use language models to create an application for inference on edge devices. The inference must have the lowest latency possible.

Which solution will meet these requirements?

A.

Deploy optimized small language models (SLMs) on edge devices.

B.

Deploy optimized large language models (LLMs) on edge devices.

C.

Incorporate a centralized small language model (SLM) API for asynchronous communication with edge devices.

D.

Incorporate a centralized large language model (LLM) API for asynchronous communication with edge devices.

Full Access
Question # 26

A company is building an application that needs to generate synthetic data that is based on existing data.

Which type of model can the company use to meet this requirement?

A.

Generative adversarial network (GAN)

B.

XGBoost

C.

Residual neural network

D.

WaveNet

Full Access