Month End Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: dumps65

Amazon Web Services AIF-C01 Dumps

Page: 1 / 14
Total 137 questions

AWS Certified AI Practitioner Exam Questions and Answers

Question 1

What are tokens in the context of generative AI models?

Options:

A.

Tokens are the basic units of input and output that a generative AI model operates on, representing words, subwords, or other linguistic units.

B.

Tokens are the mathematical representations of words or concepts used in generative AI models.

C.

Tokens are the pre-trained weights of a generative AI model that are fine-tuned for specific tasks.

D.

Tokens are the specific prompts or instructions given to a generative AI model to generate output.

Question 2

A company wants to create an application by using Amazon Bedrock. The company has a limited budget and prefers flexibility without long-term commitment.

Which Amazon Bedrock pricing model meets these requirements?

Options:

A.

On-Demand

B.

Model customization

C.

Provisioned Throughput

D.

Spot Instance

Question 3

A company needs to train an ML model to classify images of different types of animals. The company has a large dataset of labeled images and will not label more data. Which type of learning should the company use to train the model?

Options:

A.

Supervised learning.

B.

Unsupervised learning.

C.

Reinforcement learning.

D.

Active learning.

Question 4

Which technique breaks a complex task into smaller subtasks that are sent sequentially to a large language model (LLM)?

Options:

A.

One-shot prompting

B.

Prompt chaining

C.

Tree of thoughts

D.

Retrieval Augmented Generation (RAG)

Question 5

A company has a foundation model (FM) that was customized by using Amazon Bedrock to answer customer queries about products. The company wants to validate the model's responses to new types of queries. The company needs to upload a new dataset that Amazon Bedrock can use for validation.

Which AWS service meets these requirements?

Options:

A.

Amazon S3

B.

Amazon Elastic Block Store (Amazon EBS)

C.

Amazon Elastic File System (Amazon EFS)

D.

AWS Snowcone

Question 6

A company wants to use a pre-trained generative AI model to generate content for its marketing campaigns. The company needs to ensure that the generated content aligns with the company's brand voice and messaging requirements.

Which solution meets these requirements?

Options:

A.

Optimize the model's architecture and hyperparameters to improve the model's overall performance.

B.

Increase the model's complexity by adding more layers to the model's architecture.

C.

Create effective prompts that provide clear instructions and context to guide the model's generation.

D.

Select a large, diverse dataset to pre-train a new generative model.

Question 7

Why does overfilting occur in ML models?

Options:

A.

The training dataset does not reptesent all possible input values.

B.

The model contains a regularization method.

C.

The model training stops early because of an early stopping criterion.

D.

The training dataset contains too many features.

Question 8

A company has developed an ML model for image classification. The company wants to deploy the model to production so that a web application can use the model.

The company needs to implement a solution to host the model and serve predictions without managing any of the underlying infrastructure.

Which solution will meet these requirements?

Options:

A.

Use Amazon SageMaker Serverless Inference to deploy the model.

B.

Use Amazon CloudFront to deploy the model.

C.

Use Amazon API Gateway to host the model and serve predictions.

D.

Use AWS Batch to host the model and serve predictions.

Question 9

A company is using a large language model (LLM) on Amazon Bedrock to build a chatbot. The chatbot processes customer support requests. To resolve a request, the customer and the chatbot must interact a few times.

Which solution gives the LLM the ability to use content from previous customer messages?

Options:

A.

Turn on model invocation logging to collect messages.

B.

Add messages to the model prompt.

C.

Use Amazon Personalize to save conversation history.

D.

Use Provisioned Throughput for the LLM.

Question 10

An AI practitioner wants to use a foundation model (FM) to design a search application. The search application must handle queries that have text and images.

Which type of FM should the AI practitioner use to power the search application?

Options:

A.

Multi-modal embedding model

B.

Text embedding model

C.

Multi-modal generation model

D.

Image generation model

Question 11

A company wants to use generative AI to increase developer productivity and software development. The company wants to use Amazon Q Developer.

What can Amazon Q Developer do to help the company meet these requirements?

Options:

A.

Create software snippets, reference tracking, and open-source license tracking.

B.

Run an application without provisioning or managing servers.

C.

Enable voice commands for coding and providing natural language search.

D.

Convert audio files to text documents by using ML models.

Question 12

A company wants to use a large language model (LLM) on Amazon Bedrock for sentiment analysis. The company wants to classify the sentiment of text passages as positive or negative.

Which prompt engineering strategy meets these requirements?

Options:

A.

Provide examples of text passages with corresponding positive or negative labels in the prompt followed by the new text passage to be classified.

B.

Provide a detailed explanation of sentiment analysis and how LLMs work in the prompt.

C.

Provide the new text passage to be classified without any additional context or examples.

D.

Provide the new text passage with a few examples of unrelated tasks, such as text summarization or question answering.

Question 13

A company wants to build an ML model by using Amazon SageMaker. The company needs to share and manage variables for model development across multiple teams.

Which SageMaker feature meets these requirements?

Options:

A.

Amazon SageMaker Feature Store

B.

Amazon SageMaker Data Wrangler

C.

Amazon SageMaker Clarify

D.

Amazon SageMaker Model Cards

Question 14

A company has built a solution by using generative AI. The solution uses large language models (LLMs) to translate training manuals from English into other languages. The company wants to evaluate the accuracy of the solution by examining the text generated for the manuals.

Which model evaluation strategy meets these requirements?

Options:

A.

Bilingual Evaluation Understudy (BLEU)

B.

Root mean squared error (RMSE)

C.

Recall-Oriented Understudy for Gisting Evaluation (ROUGE)

D.

F1 score

Question 15

A company wants to build an ML application.

Select and order the correct steps from the following list to develop a well-architected ML workload. Each step should be selected one time. (Select and order FOUR.)

• Deploy model

• Develop model

• Monitor model

• Define business goal and frame ML problem

as

Options:

Question 16

An AI company periodically evaluates its systems and processes with the help of independent software vendors (ISVs). The company needs to receive email message notifications when an ISV's compliance reports become available.

Which AWS service can the company use to meet this requirement?

Options:

A.

AWS Audit Manager

B.

AWS Artifact

C.

AWS Trusted Advisor

D.

AWS Data Exchange

Question 17

A financial institution is building an AI solution to make loan approval decisions by using a foundation model (FM). For security and audit purposes, the company needs the AI solution's decisions to be explainable.

Which factor relates to the explainability of the AI solution's decisions?

Options:

A.

Model complexity

B.

Training time

C.

Number of hyperparameters

D.

Deployment time

Question 18

A financial institution is using Amazon Bedrock to develop an AI application. The application is hosted in a VPC. To meet regulatory compliance standards, the VPC is not allowed access to any internet traffic.

Which AWS service or feature will meet these requirements?

Options:

A.

AWS PrivateLink

B.

Amazon Macie

C.

Amazon CloudFront

D.

Internet gateway

Question 19

An AI practitioner trained a custom model on Amazon Bedrock by using a training dataset that contains confidential data. The AI practitioner wants to ensure that the custom model does not generate inference responses based on confidential data.

How should the AI practitioner prevent responses based on confidential data?

Options:

A.

Delete the custom model. Remove the confidential data from the training dataset. Retrain the custom model.

B.

Mask the confidential data in the inference responses by using dynamic data masking.

C.

Encrypt the confidential data in the inference responses by using Amazon SageMaker.

D.

Encrypt the confidential data in the custom model by using AWS Key Management Service (AWS KMS).

Question 20

A research company implemented a chatbot by using a foundation model (FM) from Amazon Bedrock. The chatbot searches for answers to questions from a large database of research papers.

After multiple prompt engineering attempts, the company notices that the FM is performing poorly because of the complex scientific terms in the research papers.

How can the company improve the performance of the chatbot?

Options:

A.

Use few-shot prompting to define how the FM can answer the questions.

B.

Use domain adaptation fine-tuning to adapt the FM to complex scientific terms.

C.

Change the FM inference parameters.

D.

Clean the research paper data to remove complex scientific terms.

Question 21

What does an F1 score measure in the context of foundation model (FM) performance?

Options:

A.

Model precision and recall

B.

Model speed in generating responses

C.

Financial cost of operating the model

D.

Energy efficiency of the model's computations

Question 22

A company is using Amazon SageMaker Studio notebooks to build and train ML models. The company stores the data in an Amazon S3 bucket. The company needs to manage the flow of data from Amazon S3 to SageMaker Studio notebooks.

Which solution will meet this requirement?

Options:

A.

Use Amazon Inspector to monitor SageMaker Studio.

B.

Use Amazon Macie to monitor SageMaker Studio.

C.

Configure SageMaker to use a VPC with an S3 endpoint.

D.

Configure SageMaker to use S3 Glacier Deep Archive.

Question 23

A pharmaceutical company wants to analyze user reviews of new medications and provide a concise overview for each medication. Which solution meets these requirements?

Options:

A.

Create a time-series forecasting model to analyze the medication reviews by using Amazon Personalize.

B.

Create medication review summaries by using Amazon Bedrock large language models (LLMs).

C.

Create a classification model that categorizes medications into different groups by using Amazon SageMaker.

D.

Create medication review summaries by using Amazon Rekognition.

Question 24

A company wants to build a lead prioritization application for its employees to contact potential customers. The application must give employees the ability to view and adjust the weights assigned to different variables in the model based on domain knowledge and expertise.

Which ML model type meets these requirements?

Options:

A.

Logistic regression model

B.

Deep learning model built on principal components

C.

K-nearest neighbors (k-NN) model

D.

Neural network

Question 25

An ecommerce company is deploying a chatbot. The chatbot will give users the ability to ask questions about the company's products and receive details on users' orders. The company must implement safeguards for the chatbot to filter harmful content from the input prompts and chatbot responses.

Which AWS feature or resource meets these requirements?

Options:

A.

Amazon Bedrock Guardrails

B.

Amazon Bedrock Agents

C.

Amazon Bedrock inference APIs

D.

Amazon Bedrock custom models

Question 26

A company is using an Amazon Bedrock base model to summarize documents for an internal use case. The company trained a custom model to improve the summarization quality.

Which action must the company take to use the custom model through Amazon Bedrock?

Options:

A.

Purchase Provisioned Throughput for the custom model.

B.

Deploy the custom model in an Amazon SageMaker endpoint for real-time inference.

C.

Register the model with the Amazon SageMaker Model Registry.

D.

Grant access to the custom model in Amazon Bedrock.

Question 27

A company needs to choose a model from Amazon Bedrock to use internally. The company must identify a model that generates responses in a style that the company's employees prefer.

What should the company do to meet these requirements?

Options:

A.

Evaluate the models by using built-in prompt datasets.

B.

Evaluate the models by using a human workforce and custom prompt datasets.

C.

Use public model leaderboards to identify the model.

D.

Use the model InvocationLatency runtime metrics in Amazon CloudWatch when trying models.

Question 28

A company is using a pre-trained large language model (LLM) to extract information from documents. The company noticed that a newer LLM from a different provider is available on Amazon Bedrock. The company wants to transition to the new LLM on Amazon Bedrock.

What does the company need to do to transition to the new LLM?

Options:

A.

Create a new labeled dataset

B.

Perform feature engineering.

C.

Adjust the prompt template.

D.

Fine-tune the LLM.

Question 29

A loan company is building a generative AI-based solution to offer new applicants discounts based on specific business criteria. The company wants to build and use an AI model responsibly to minimize bias that could negatively affect some customers.

Which actions should the company take to meet these requirements? (Select TWO.)

Options:

A.

Detect imbalances or disparities in the data.

B.

Ensure that the model runs frequently.

C.

Evaluate the model's behavior so that the company can provide transparency to stakeholders.

D.

Use the Recall-Oriented Understudy for Gisting Evaluation (ROUGE) technique to ensure that the model is 100% accurate.

E.

Ensure that the model's inference time is within the accepted limits.

Question 30

A company is using the Generative AI Security Scoping Matrix to assess security responsibilities for its solutions. The company has identified four different solution scopes based on the matrix.

Which solution scope gives the company the MOST ownership of security responsibilities?

Options:

A.

Using a third-party enterprise application that has embedded generative AI features.

B.

Building an application by using an existing third-party generative AI foundation model (FM).

C.

Refining an existing third-party generative AI foundation model (FM) by fine-tuning the model by using data specific to the business.

D.

Building and training a generative AI model from scratch by using specific data that a customer owns.

Question 31

Which technique involves training AI models on labeled datasets to adapt the models to specific industry terminology and requirements?

Options:

A.

Data augmentation

B.

Fine-tuning

C.

Model quantization

D.

Continuous pre-training

Question 32

A company uses a foundation model (FM) from Amazon Bedrock for an AI search tool. The company wants to fine-tune the model to be more accurate by using the company's data.

Which strategy will successfully fine-tune the model?

Options:

A.

Provide labeled data with the prompt field and the completion field.

B.

Prepare the training dataset by creating a .txt file that contains multiple lines in .csv format.

C.

Purchase Provisioned Throughput for Amazon Bedrock.

D.

Train the model on journals and textbooks.

Question 33

A digital devices company wants to predict customer demand for memory hardware. The company does not have coding experience or knowledge of ML algorithms and needs to develop a data-driven predictive model. The company needs to perform analysis on internal data and external data.

Which solution will meet these requirements?

Options:

A.

Store the data in Amazon S3. Create ML models and demand forecast predictions by using Amazon SageMaker built-in algorithms that use the data from Amazon S3.

B.

Import the data into Amazon SageMaker Data Wrangler. Create ML models and demand forecast predictions by using SageMaker built-in algorithms.

C.

Import the data into Amazon SageMaker Data Wrangler. Build ML models and demand forecast predictions by using an Amazon Personalize Trending-Now recipe.

D.

Import the data into Amazon SageMaker Canvas. Build ML models and demand forecast predictions by selecting the values in the data from SageMaker Canvas.

Question 34

A company has thousands of customer support interactions per day and wants to analyze these interactions to identify frequently asked questions and develop insights.

Which AWS service can the company use to meet this requirement?

Options:

A.

Amazon Lex

B.

Amazon Comprehend

C.

Amazon Transcribe

D.

Amazon Translate

Question 35

A company's large language model (LLM) is experiencing hallucinations.

How can the company decrease hallucinations?

Options:

A.

Set up Agents for Amazon Bedrock to supervise the model training.

B.

Use data pre-processing and remove any data that causes hallucinations.

C.

Decrease the temperature inference parameter for the model.

D.

Use a foundation model (FM) that is trained to not hallucinate.

Question 36

An ecommerce company is using a chatbot to automate the customer order submission process. The chatbot is powered by AI and Is available to customers directly from the company's website 24 hours a day, 7 days a week.

Which option is an AI system input vulnerability that the company needs to resolve before the chatbot is made available?

Options:

A.

Data leakage

B.

Prompt injection

C.

Large language model (LLM) hallucinations

D.

Concept drift

Question 37

A company has built a chatbot that can respond to natural language questions with images. The company wants to ensure that the chatbot does not return inappropriate or unwanted images.

Which solution will meet these requirements?

Options:

A.

Implement moderation APIs.

B.

Retrain the model with a general public dataset.

C.

Perform model validation.

D.

Automate user feedback integration.

Question 38

A company wants to classify human genes into 20 categories based on gene characteristics. The company needs an ML algorithm to document how the inner mechanism of the model affects the output.

Which ML algorithm meets these requirements?

Options:

A.

Decision trees

B.

Linear regression

C.

Logistic regression

D.

Neural networks

Question 39

A company built an AI-powered resume screening system. The company used a large dataset to train the model. The dataset contained resumes that were not representative of all demographics. Which core dimension of responsible AI does this scenario present?

Options:

A.

Fairness.

B.

Explainability.

C.

Privacy and security.

D.

Transparency.

Question 40

A company wants to develop an Al application to help its employees check open customer claims, identify details for a specific claim, and access documents for a claim. Which solution meets these requirements?

Options:

A.

Use Agents for Amazon Bedrock with Amazon Fraud Detector to build the application.

B.

Use Agents for Amazon Bedrock with Amazon Bedrock knowledge bases to build the application.

C.

Use Amazon Personalize with Amazon Bedrock knowledge bases to build the application.

D.

Use Amazon SageMaker AI to build the application by training a new ML model.

Question 41

A company wants to develop an educational game where users answer questions such as the following: "A jar contains six red, four green, and three yellow marbles. What is the probability of choosing a green marble from the jar?"

Which solution meets these requirements with the LEAST operational overhead?

Options:

A.

Use supervised learning to create a regression model that will predict probability.

B.

Use reinforcement learning to train a model to return the probability.

C.

Use code that will calculate probability by using simple rules and computations.

D.

Use unsupervised learning to create a model that will estimate probability density.

Page: 1 / 14
Total 137 questions