A company is developing a generative AI (GenAI) application that uses Amazon Bedrock
foundation models. The application has several custom tool integrations. The application
has experienced unexpected token consumption surges despite consistent user traffic.
The company needs a solution that uses Amazon Bedrock model invocation logging to
monitor InputTokenCount and OutputTokenCount metrics. The solution must detect
unusual patterns in tool usage and identify which specific tool integrations cause abnormal
token consumption. The solution must also automatically adjust thresholds as traffic
patterns change.
Which solution will meet these requirements?
A. Use Amazon CloudWatch Logs to capture model invocation logs. Create CloudWatch dashboards for token metrics. Configure static CloudWatch alarms with fixed thresholds for each tool integration.
B. Store model invocation logs in Amazon S3. Use AWS Glue and Amazon Athena to analyze token usage trends.
C. Use Amazon CloudWatch Logs to capture model invocation logs. Create CloudWatch metric filters to extract tool-specific invocation patterns. Apply CloudWatch anomaly detection alarms that automatically adjust baselines for each tool’s token metrics.
D. Store model invocation logs in an Amazon S3 bucket. Use AWS Lambda to process logs in real time. Manually update CloudWatch alarm thresholds based on trends identified by the Lambda function.
An ecommerce company is developing a generative AI application that uses Amazon
Bedrock with Anthropic Claude to recommend products to customers. Customers report
that some recommended products are not available for sale on the website or are not
relevant to the customer. Customers also report that the solution takes a long time to
generate some recommendations.
The company investigates the issues and finds that most interactions between customers
and the product recommendation solution are unique. The company confirms that the
solution recommends products that are not in the company’s product catalog. The
company must resolve these issues.
Which solution will meet this requirement?
A. Increase grounding within Amazon Bedrock Guardrails. Enable Automated Reasoning checks. Set up provisioned throughput.
B. Use prompt engineering to restrict the model responses to relevant products. Use streaming techniques such as the InvokeModelWithResponseStream action to reduce perceived latency for the customers.
C. Create an Amazon Bedrock knowledge base. Implement Retrieval Augmented Generation RAG. Set the PerformanceConfigLatency parameter to optimized.
D. Store product catalog data in Amazon OpenSearch Service. Validate the model’s product recommendations against the product catalog. Use Amazon DynamoDB to implement response caching.
A company runs a Retrieval Augmented Generation (RAG) application that uses Amazon
Bedrock Knowledge Bases to perform regulatory compliance queries. The application uses
the RetrieveAndGenerateStream API. The application retrieves relevant documents from a
knowledge base that contains more than 50,000 regulatory documents, legal precedents,
and policy updates.
The RAG application is producing suboptimal responses because the initial retrieval often
returns semantically similar but contextually irrelevant documents. The poor responses are
causing model hallucinations and incorrect regulatory guidance. The company needs to
improve the performance of the RAG application so it returns more relevant documents.
Which solution will meet this requirement with the LEAST operational overhead?
A. Deploy an Amazon SageMaker endpoint to run a fine-tuned ranking model. Use an Amazon API Gateway REST API to route requests. Configure the application to make requests through the REST API to rerank the results.
B. Use Amazon Comprehend to classify documents and apply relevance scores. Integrate the RAG application’s reranking process with Amazon Textract to run document analysis. Use Amazon Neptune to perform graph-based relevance calculations.
C. Implement a retrieval pipeline that uses the Amazon Bedrock Knowledge Bases Retrieve API to perform initial document retrieval. Call the Amazon Bedrock Rerank API to rerank the results. Invoke the InvokeModelWithResponseStream operation to generate responses.
D. Use the latest Amazon reranker model through the reranking configuration within Amazon Bedrock Knowledge Bases. Use the model to improve document relevance scoring and to reorder results based on contextual assessments.
A company is building a serverless application that uses AWS Lambda functions to help
students around the world summarize notes. The application uses Anthropic Claude
through Amazon Bedrock. The company observes that most of the traffic occurs during
evenings in each time zone. Users report experiencing throttling errors during peak usage
times in their time zones.
The company needs to resolve the throttling issues by ensuring continuous operation of the
application. The solution must maintain application performance quality and must not
require a fixed hourly cost during low traffic periods.
Which solution will meet these requirements?
A. Create custom Amazon CloudWatch metrics to monitor model errors. Set provisioned throughput to a value that is safely higher than the peak traffic observed.
B. Create custom Amazon CloudWatch metrics to monitor model errors. Set up a failover mechanism to redirect invocations to a backup AWS Region when the errors exceed a specified threshold.
C. Enable invocation logging in Amazon Bedrock. Monitor key metrics such as Invocations, InputTokenCount, OutputTokenCount, and InvocationThrottles. Distribute traffic across cross-Region inference endpoints.
D. Enable invocation logging in Amazon Bedrock. Monitor InvocationLatency, InvocationClientErrors, and InvocationServerErrors metrics. Distribute traffic across multiple versions of the same model.
A healthcare company is using Amazon Bedrock to develop a real-time patient care AI
assistant to respond to queries for separate departments that handle clinical inquiries,
insurance verification, appointment scheduling, and insurance claims. The company wants
to use a multi-agent architecture.
The company must ensure that the AI assistant is scalable and can onboard new features
for patients. The AI assistant must be able to handle thousands of parallel patient
interactions. The company must ensure that patients receive appropriate domain-specific
responses to queries.
Which solution will meet these requirements?
A. Isolate data for each agent by using separate knowledge bases. Use IAM filtering to control access to each knowledge base. Deploy a supervisor agent to perform natural language intent classification on patient inquiries. Configure the supervisor agent to route queries to specialized collaborator agents to respond to department-specific queries. Configure each specialized collaborator agent to use Retrieval Augmented Generation (RAG) with the agent’s department-specific knowledge base.
B. Create a separate supervisor agent for each department. Configure individual collaborator agents to perform natural language intent classification for each specialty domain within each department. Integrate each collaborator agent with department-specific knowledge bases only. Implement manual handoff processes between the supervisor agents.
C. Isolate data for each department in separate knowledge bases. Use IAM filtering to control access to each knowledge base. Deploy a single general-purpose agent. Configure multiple action groups within the general-purpose agent to perform specific department functions. Implement rule-based routing logic within the general-purpose agent instructions.
D. Implement multiple independent supervisor agents that run in parallel to respond to patient inquiries for each department. Configure multiple collaborator agents for each supervisor agent. Integrate all agents with the same knowledge base. Use external routing logic to merge responses from multiple supervisor agents.
A company has a generative AI (GenAI) application that uses Amazon Bedrock to provide
real-time responses to customer queries. The company has noticed intermittent failures
with API calls to foundation models (FMs) during peak traffic periods.
The company needs a solution to handle transient errors and provide detailed observability
into FM performance. The solution must prevent cascading failures during throttling events
and provide distributed tracing across service boundaries to identify latency contributors.
The solution must also enable correlation of performance issues with specific FM
characteristics.
Which solution will meet these requirements?
A. Implement a custom retry mechanism with a fixed delay of 1 second between retries. Configure Amazon CloudWatch alarms to monitor the application’s error rates and latency metrics.
B. Configure the AWS SDK with standard retry mode and exponential backoff with jitter. Use AWS X-Ray tracing with annotations to identify and filter service components.
C. Implement client-side caching of all FM responses. Add custom logging statements in the application code to record API call durations.
D. Configure the AWS SDK with adaptive retry mode. Use AWS CloudTrail distributed tracing to monitor throttling events.
A company deploys multiple Amazon Bedrock–based generative AI (GenAI) applications
across multiple business units for customer service, content generation, and document
analysis. Some applications show unpredictable token consumption patterns. The company
requires a comprehensive observability solution that provides real-time visibility into token
usage patterns across multiple models. The observability solution must support custom
dashboards for multiple stakeholder groups and provide alerting capabilities for token
consumption across all the foundation models that the company’s applications use.
Which combination of solutions will meet these requirements with the LEAST operational
overhead? (Select TWO.)
A. Use Amazon CloudWatch metrics as data sources to create custom Amazon QuickSight dashboards that show token usage trends and usage patterns across FMs.
B. Use CloudWatch Logs Insights to analyze Amazon Bedrock invocation logs for token consumption patterns and usage attribution by application. Create custom queries to identify high-usage scenarios. Add log widgets to dashboards to enable continuous monitoring.
C. Create custom Amazon CloudWatch dashboards that combine native Amazon Bedrock token and invocation CloudWatch metrics. Set up CloudWatch alarms to monitor token usage thresholds.
D. Create dashboards that show token usage trends and patterns across the company’s FMs by using an Amazon Bedrock zero-ETL integration with Amazon Managed Grafana.
E. Implement Amazon EventBridge rules to capture Amazon Bedrock model invocation events. Route token usage data to Amazon OpenSearch Serverless by using Amazon Data Firehose. Use OpenSearch dashboards to analyze usage patterns.
A retail company is using Amazon Bedrock to develop a customer service AI assistant.
Analysis shows that 70% of customer inquiries are simple product questions that a smaller
model can effectively handle. However, 30% of inquiries are complex return policy
questions that require advanced reasoning.
The company wants to implement a cost-effective model selection framework to automatically route customer inquiries to appropriate models based on inquiry complexity.
The framework must maintain high customer satisfaction and minimize response latency.
Which solution will meet these requirements with the LEAST implementation effort?
A. Create a multi-stage architecture that uses a small foundation model (FM) to classify the complexity of each inquiry. Route simple inquiries to a smaller, more cost-effective model. Route complex inquiries to a larger, more capable model. Use AWS Lambda functions to handle routing logic.
B. Use Amazon Bedrock intelligent prompt routing to automatically analyze inquiries. Route simple product inquiries to smaller models and route complex return policy inquiries to more capable larger models.
C. Implement a single-model solution that uses an Amazon Bedrock mid-sized foundation model (FM) with on-demand pricing. Include special instructions in model prompts to handle both simple and complex inquiries by using the same model.
D. Create separate Amazon Bedrock endpoints for simple and complex inquiries. Implement a rule-based routing system based on keyword detection. Use on-demand pricing for the smaller model and provisioned throughput for the larger model.
A financial services company is deploying a generative AI (GenAI) application that uses
Amazon Bedrock to assist customer service representatives to provide personalized
investment advice to customers. The company must implement a comprehensive
governance solution that follows responsible AI practices and meets regulatory
requirements.
The solution must detect and prevent hallucinations in recommendations. The solution
must have safety controls for customer interactions. The solution must also monitor model
behavior drift in real time and maintain audit trails of all prompt-response pairs for
regulatory review. The company must deploy the solution within 60 days. The solution must
integrate with the company's existing compliance dashboard and respond to customers
within 200 ms.
Which solution will meet these requirements with the LEAST operational overhead?
A. Configure Amazon Bedrock guardrails to apply custom content filters and toxicity detection. Use Amazon Bedrock Model Evaluation to detect hallucinations. Store promptresponse pairs in Amazon DynamoDB to capture audit trails and set a TTL. Integrate Amazon CloudWatch custom metrics with the existing compliance dashboard.
B. Deploy Amazon Bedrock and use AWS PrivateLink to access the application securely. Use AWS Lambda functions to implement custom prompt validation. Store prompt-response pairs in an Amazon S3 bucket and configure S3 Lifecycle policies. Create custom Amazon CloudWatch dashboards to monitor model performance metrics.
C. Use Amazon Bedrock Agents and Amazon Bedrock Knowledge Bases to ground responses. Use Amazon Bedrock Guardrails to enforce content safety. Use Amazon OpenSearch Service to store and index prompt-response pairs. Integrate OpenSearch Service with Amazon QuickSight to create compliance reports and to detect model behavior drift.
D. Use Amazon SageMaker Model Monitor to detect model behavior drift. Use AWS WAF to filter content. Store customer interactions in an encrypted Amazon RDS database. Use Amazon API Gateway to create custom HTTP APIs to integrate with the compliance dashboard.
A company is building a legal research AI assistant that uses Amazon Bedrock with an
Anthropic Claude foundation model (FM). The AI assistant must retrieve highly relevant
case law documents to augment the FM’s responses. The AI assistant must identify
semantic relationships between legal concepts, specific legal terminology, and citations.
The AI assistant must perform quickly and return precise results.
Which solution will meet these requirements?
A. Configure an Amazon Bedrock knowledge base to use a default vector search configuration. Use Amazon Bedrock to expand queries to improve retrieval for legal documents based on specific terminology and citations.
B. Use Amazon OpenSearch Service to deploy a hybrid search architecture that combines vector search with keyword search. Apply an Amazon Bedrock reranker model to optimize result relevance.
C. Enable the Amazon Kendra query suggestion feature for end users. Use Amazon Bedrock to perform post-processing of search results to identify semantic similarity in the documents and to produce precise results.
D. Use Amazon OpenSearch Service with vector search and Amazon Bedrock Titan Embeddings to index and search legal documents. Use custom AWS Lambda functions to merge results with keyword-based filters that are stored in an Amazon RDS database.
A company uses an AI assistant application to summarize the company’s website content
and provide information to customers. The company plans to use Amazon Bedrock to give
the application access to a foundation model (FM).
The company needs to deploy the AI assistant application to a development environment
and a production environment. The solution must integrate the environments with the FM.
The company wants to test the effectiveness of various FMs in each environment. The
solution must provide product owners with the ability to easily switch between FMs for
testing purposes in each environment.
Which solution will meet these requirements?
A. Create one AWS CDK application. Create multiple pipelines in AWS CodePipeline. Configure each pipeline to have its own settings for each FM. Configure the application to invoke the Amazon Bedrock FMs by using the aws_bedrock.ProvisionedModel.fromProvisionedModelArn() method.
B. Create a separate AWS CDK application for each environment. Configure the applications to invoke the Amazon Bedrock FMs by using the aws_bedrock.FoundationModel.fromFoundationModelId() method. Create a separate pipeline in AWS CodePipeline for each environment.
C. Create one AWS CDK application. Configure the application to invoke the Amazon Bedrock FMs by using the aws_bedrock.FoundationModel.fromFoundationModelId() method. Create a pipeline in AWS CodePipeline that has a deployment stage for each environment that uses AWS CodeBuild deploy actions.
D. Create one AWS CDK application for the production environment. Configure the application to invoke the Amazon Bedrock FMs by using the aws_bedrock.ProvisionedModel.fromProvisionedModelArn() method. Create a pipeline in AWS CodePipeline. Configure the pipeline to deploy to the production environment by using an AWS CodeBuild deploy action. For the development environment, manually recreate the resources by referring to the production application code.
A GenAI developer is building a Retrieval Augmented Generation (RAG)-based customer
support application that uses Amazon Bedrock foundation models (FMs). The application
needs to process 50 GB of historical customer conversations that are stored in an Amazon
S3 bucket as JSON files. The application must use the processed data as its retrieval
corpus. The application’s data processing workflow must extract relevant data from
customer support documents, remove customer personally identifiable information (PII),
and generate embeddings for vector storage. The processing workflow must be costeffective
and must finish within 4 hours.
Which solution will meet these requirements with the LEAST operational overhead?
A. Use AWS Lambda and Amazon Comprehend to process files in parallel, remove PII, and call Amazon Bedrock APIs to generate vectors. Configure Lambda concurrency limits and memory settings to optimize throughput.
B. Create an AWS Glue ETL job to run PII detection scripts on the data. Use Amazon SageMaker Processing to run the HuggingFaceProcessor to generate embeddings by using a pre-trained model. Store the embeddings in Amazon OpenSearch Service.
C. Deploy an Amazon EMR cluster that runs Apache Spark with user-defined functions (UDFs) that call Amazon Comprehend to detect PII. Use Amazon Bedrock APIs to generate vectors. Store outputs in Amazon Aurora PostgreSQL with the pgvector extension.
D. Implement a data processing pipeline that uses AWS Step Functions to orchestrate a workload that uses Amazon Comprehend to detect PII and Amazon Bedrock to generate embeddings. Directly integrate the workflow with Amazon OpenSearch Serverless to store vectors and provide similarity search capabilities.
| Page 1 out of 9 Pages |
| 123 |
Real-World Scenario Mastery: Our AIP-C01 practice exam don't just test definitions. They present you with the same complex, scenario-based problems you'll encounter on the actual exam.
Strategic Weakness Identification: Each practice session reveals exactly where you stand. Discover which domains need more attention, before AWS Certified Generative AI Developer - Professional exam day arrives.
Confidence Through Familiarity: There's no substitute for knowing what to expect. When you've worked through our comprehensive AIP-C01 practice exam questions pool covering all topics, the real exam feels like just another practice session.