Spring Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70track

Free Amazon Web Services DVA-C02 Practice Exam with Questions & Answers | Set: 4

Questions 46

A developer maintains a serverless application that uses AWS Lambda to process financial transaction files uploaded to an Amazon S3 bucket. A Lambda function is scheduled to run once per hour to process the files.

As file uploads increase, the Lambda function occasionally exceeds the 15-minute execution limit . The developer must handle higher file volumes, reduce processing time, and prevent duplicate file processing .

Which solution will meet these requirements?

Options:
A.

Configure Amazon S3 event notifications to invoke the Lambda function for each uploaded file. Track processed file identifiers in Amazon DynamoDB.

B.

Move the application logic to a single Amazon EC2 instance to process files more efficiently.

C.

Use Amazon EventBridge Scheduler to run the Lambda function every 5 minutes to scan the S3 bucket.

D.

Continuously poll the S3 bucket from Lambda and rely on each object’s LastModified timestamp.

Amazon Web Services DVA-C02 Premium Access
Questions 47

A company is using Amazon RDS as the Backend database for its application. After a recent marketing campaign, a surge of read requests to the database increased the latency of data retrieval from the database.

The company has decided to implement a caching layer in front of the database. The cached content must be encrypted and must be highly available.

Which solution will meet these requirements?

Options:
A.

Amazon Cloudfront

B.

Amazon ElastiCache to Memcached

C.

Amazon ElastiCache for Redis in cluster mode

D.

Amazon DynamoDB Accelerate (DAX)

Questions 48

A company has deployed infrastructure on AWS. A development team wants to create an AWS Lambda function that will retrieve data from an Amazon Aurora database. The Amazon Aurora database is in a private subnet in company ' s VPC. The VPC is named VPC1. The data is relational in nature. The Lambda function needs to access the data securely.

Which solution will meet these requirements?

Options:
A.

Create the Lambda function. Configure VPC1 access for the function. Attach a security group named SG1 to both the Lambda function and the database. Configure the security group inbound and outbound rules to allow TCP traffic on Port 3306.

B.

Create and launch a Lambda function in a new public subnet that is in a new VPC named VPC2. Create a peering connection between VPC1 and VPC2.

C.

Create the Lambda function. Configure VPC1 access for the function. Assign a security group named SG1 to the Lambda function. Assign a second security group named SG2 to the database. Add an inbound rule to SG1 to allow TCP traffic from Port 3306.

D.

Export the data from the Aurora database to Amazon S3. Create and launch a Lambda function in VPC1. Configure the Lambda function query the data from Amazon S3.

Questions 49

A developer designed an application on an Amazon EC2 instance The application makes API requests to objects in an Amazon S3 bucket

Which combination of steps will ensure that the application makes the API requests in the MOST secure manner? (Select TWO.)

Options:
A.

Create an IAM user that has permissions to the S3 bucket. Add the user to an 1AM group

B.

Create an IAM role that has permissions to the S3 bucket

C.

Add the IAM role to an instance profile. Attach the instance profile to the EC2 instance.

D.

Create an 1AM role that has permissions to the S3 bucket Assign the role to an 1AM group

E.

Store the credentials of the IAM user in the environment variables on the EC2 instance

Questions 50

A company wants to automate part of its deployment process. A developer needs to automate the process of checking for and deleting unused resources that supported previously deployed stacks but that are no longer used.

The company has a central application that uses the AWS Cloud Development Kit (AWS CDK) to manage all deployment stacks. The stacks are spread out across multiple accounts. The developer’s solution must integrate as seamlessly as possible within the current deployment process.

Which solution will meet these requirements with the LEAST amount of configuration?

Options:
A.

In the central AWS CDK application, write a handler function in the code that uses AWS SDK calls to check for and delete unused resources. Create an AWS CloudPormation template from a JSON file. Use the template to attach the function code to an AWS Lambda function and lo invoke the Lambda function when the deployment slack runs.

B.

In the central AWS CDK application. write a handler function in the code that uses AWS SDK calls to check for and delete unused resources. Create an AWS CDK custom resource Use the custom resource to attach the function code to an AWS Lambda function and to invoke the Lambda function when the deployment stack runs.

C.

In the central AWS CDK, write a handler function m the code that uses AWS SDK calls to check for and delete unused resources. Create an API in AWS Amplify Use the API to attach the function code to an AWS Lambda function and to invoke the Lambda function when the deployment stack runs.

D.

In the AWS Lambda console write a handler function in the code that uses AWS SDK calls to check for and delete unused resources. Create an AWS CDK custom resource. Use the custom resource to import the Lambda function into the stack and to Invoke the Lambda function when the deployment stack runs.

Questions 51

A developer is building an application that stores objects in an Amazon S3 bucket. The bucket does not have versioning enabled. The objects are accessed rarely after 1 week. However, the objects must be immediately available at all times.

The developer wants to optimize storage costs for the S3 bucket.

Which solution will meet this requirement?

Options:
A.

Create an S3 Lifecycle rule to expire objects after 7 days.

B.

Create an S3 Lifecycle rule to transition objects to S3 Standard-Infrequent Access (S3 Standard-IA) after 7 days.

C.

Create an S3 Lifecycle rule to transition objects to S3 Glacier Flexible Retrieval after 7 days.

D.

Create an S3 Lifecycle rule to delete objects that have delete markers.

Questions 52

A developer writes an AWS Lambda function that processes new object uploads to an Amazon S3 bucket. The Lambda function runs for approximately 30 seconds. The function runs as expected under normal load conditions. Other Lambda functions in the AWS account also run as expected. Occasionally, up to 500 new objects are written to the bucket every minute. Each new object write invokes the processing Lambda function during the high-volume periods through an event. The developer must ensure that the processing function will continue to run as expected during the high-volume periods. Which solution will meet this requirement?

Options:
A.

Modify the function ' s timeout setting.

B.

Add an additional layer to optimize the code execution.

C.

Configure a reserved concurrency quota for the function.

D.

Decrease the function’s memory allocation.

Questions 53

A developer is building a serverless application that uses asynchronous AWS Lambda functions. The developer needs a solution to capture records of every Lambda function invocation. Each function must have multiple destinations based on whether each invocation is successful. The solution must record function responses in JSON format.

Which solution will meet these requirements?

Options:
A.

Amazon CloudWatch Logs log groups that use the default log format for Lambda functions. Route each invocation to the appropriate log group by using a Lambda canary deployment and weighted aliases. Set the appropriate log group as the target for each function.

B.

Set up an S3 bucket as an on-failure destination for the Lambda function. Configure an Amazon SNS topic as the destination for successful Lambda function invocations.

C.

Configure an Amazon SQS dead-letter queue as an event source for the Lambda function to store failed invocations. In the Lambda function code, use the PutItem Amazon DynamoDB API call to add the successful invocation information to the database.

D.

Set up an Amazon SQS queue as an on-failure destination for the Lambda function. Configure an Amazon OpenSearch Service cluster as the destination for the Lambda function for successful invocations.

Questions 54

A company runs a web application on Amazon EC2 instances behind an Application Load Balancer. The application uses Amazon DynamoDB as its database. The company wants to ensure high performance for reads and writes.

Which solution will meet these requirements MOST cost-effectively?

Options:
A.

Configure auto-scaling for the DynamoDB table with a target utilization of 70%. Set the minimum and maximum capacity units based on the expected workload.

B.

Use DynamoDB on-demand capacity mode for the table. Specify a maximum throughput higher than the expected peak read and write capacity units.

C.

Use DynamoDB provisioned throughput mode for the table. Create an Amazon CloudWatch alarm on the ThrottledRequests metric. Invoke an AWS Lambda function to increase provisioned capacity.

D.

Create an Amazon DynamoDB Accelerator (DAX) cluster. Configure the application to use the DAX endpoint.

Questions 55

A developer created several AWS Lambda functions that write data to a single Amazon S3 bucket. The developer configured all the Lambda functions to send logs and metrics to Amazon CloudWatch.

The developer receives reports that one of the Lambda functions writes data to the bucket very slowly. The developer needs to measure the latency between the problematic Lambda function and the S3 bucket.

Which solution will meet this requirement?

Options:
A.

Enable AWS X-Ray on the Lambda function. In the generated trace map. select the line between Lambda and Amazon S3.

B.

Query the Lambda function ' s log file in Amazon CloudWatch Logs Insights. Return the average of the auto-discovered ©duration field.

C.

Enable CloudWatch Lambda Insights on the function. View the latency graph that CloudWatch Lambda Insights provides.

D.

Enable AWS X-Ray on the Lambda function. Select Amazon S3 in the latency graph to view the latency histogram.

Questions 56

A developer is building a multi-tenant application using AWS Lambda, Amazon S3, and Amazon DynamoDB. Each S3 object prefix represents a tenant name, and DynamoDB uses the tenant name as the partition key.

The developer must prevent cross-tenant data access during processing.

Which combination of actions will meet this requirement? (Select THREE.)

Options:
A.

Create a data access IAM role that allows the sts:TagSession action.

B.

Allow the Lambda execution role to assume the data access role.

C.

Configure IAM policies on the data access role to allow S3 and DynamoDB access only when resource attributes match the tenant session tag.

D.

Create a resource-based policy on DynamoDB based on principal tags.

E.

Create a resource control policy (RCP) for the S3 bucket.

F.

Configure the Lambda function to assume the data access role and pass the tenant name as a session tag.

Questions 57

A developer is building an application that processes a stream of user-supplied data. The data stream must be consumed by multiple Amazon EC2 based processing applications in parallel and in real time. Each processor must be able to resume without losing data if there is a service interruption. The application architect plans to add other processors in the near future, and wants to minimize the amount of data duplication involved.

Which solution will satisfy these requirements?

Options:
A.

Publish the data to Amazon Simple Queue Service (Amazon SQS).

B.

Publish the data to Amazon Data Firehose.

C.

Publish the data to Amazon EventBridge.

D.

Publish the data to Amazon Kinesis Data Streams.

Questions 58

A company is building an application to accept data from customers. The data must be encrypted at rest and in transit.

The application uses an Amazon API Gateway API that resolves to AWS Lambda functions. The Lambda functions store the data in an Amazon Aurora MySQL DB cluster. The application worked properly during testing.

A developer configured an Amazon CloudFront distribution with field-level encryption that uses an AWS Key Management Service (AWS KMS) key. After the configuration of the distribution, the application behaved unexpectedly. All the data in the database changed from plaintext to ciphertext.

The developer must ensure that the data is not stored in the database as the ciphertext from the CloudFront field-level encryption.

Which solution will meet this requirement?

Options:
A.

Change the CloudFront Viewer protocol policy from " HTTP and HTTPS " to " HTTPS only. "

B.

Add a Lambda function that uses the KMS key to decrypt the data fields before saving the data to the database.

C.

Enable encryption on the DB cluster by using the same KMS key that is used in CloudFront.

D.

Request and deploy a new SSL certificate to use with the CloudFront distribution.

Questions 59

A developer manages a website that distributes its content by using Amazon CloudFront. The website ' s static artifacts are stored in an Amazon S3 bucket.

The developer deploys some changes and can see the new artifacts in the S3 bucket. However, the changes do not appear on the webpage that the CloudFront distribution delivers.

How should the developer resolve this issue?

Options:
A.

Configure S3 Object Lock to update to the latest version of the files every time an S3 object is updated.

B.

Configure the S3 bucket to clear all old objects from the bucket before new artifacts are uploaded.

C.

Set CloudFront to invalidate the cache after the artifacts have been deployed to Amazon S3.

D.

Set CloudFront to modify the distribution origin after the artifacts have been deployed to Amazon S3.

Questions 60

A developer is updating an Amazon API Gateway REST API to have a mock endpoint. The developer wants to update the integration request mapping template so the endpoint will respond to mock integration requests with specific HTTP status codes based on various conditions.

Options:
A.

{ if( $input.params(‘integration’) == " mock " ) " statusCode " : 404 else " statusCode " : 500 end }

B.

{ if( $input.params( ' scope ' ) == " internal " ) " statusCode " : 200 else " statusCode " : 500 end }

C.

{ if( $input.path( " integration " ) ) " statusCode " : 200 else " statusCode " :404 end }

D.

{ if( $context.integration.status ) " statusCode " : 200 else " statusCode " : 500 end }