AWS DAS-C01 Free Practice Questions — Page 1

Data Analytics - Specialty • 5 questions • Answers & explanations included

Question 1

A manufacturing company wants to create an operational analytics dashboard to visualize metrics from equipment in near- real time. The company uses Amazon Kinesis Data Streams to stream the data to other applications. The dashboard must automatically refresh every 5 seconds. A data analytics specialist must design a solution that requires the least possible implementation effort.Which solution meets these requirements?

A. Use Amazon Kinesis Data Firehose to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard.
B. Use Apache Spark Streaming on Amazon EMR to read the data in near-real time. Develop a custom application for the dashboard by using D3.js.
C. Use Amazon Kinesis Data Firehose to push the data into an Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster. Visualize the data by using an OpenSearch Dashboards (Kibana).
D. Use AWS Glue streaming ETL to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard
Show Answer & Explanation

Correct Answer: C. Use Amazon Kinesis Data Firehose to push the data into an Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster. Visualize the data by using an OpenSearch Dashboards (Kibana).

Why C is correct: Amazon Kinesis Data Firehose can directly push data into Amazon OpenSearch Service (formerly Elasticsearch Service), and OpenSearch Dashboards (formerly Kibana) provides native real-time visualization capabilities with automatic refresh intervals as low as 5 seconds. This is a fully managed solution requiring minimal implementation effort - you simply configure Firehose to deliver to OpenSearch and create dashboards in the OpenSearch Dashboards UI. Why others are wrong: A & D use QuickSight, which requires manual refresh and doesn't support automatic 5-second refresh intervals for streaming data B requires significant custom development effort with D3.js and managing EMR clusters, which violates the "least implementation effort" requirement

Question 2

An IoT company wants to release a new device that will collect data to track sleep overnight on an intelligent mattress. Sensors will send data that will be uploaded to an Amazon S3 bucket. About 2 MB of data is generated each night for each bed. Data must be processed and summarized for each user, and the results need to be available as soon as possible. Part of the process consists of time windowing and other functions. Based on tests with a Python script, every run will require about 1 GB of memory and will complete within a couple of minutes.Which solution will run the script in the MOST cost-effective way?

A. AWS Lambda with a Python script
B. AWS Glue with a Scala job
C. Amazon EMR with an Apache Spark script
D. AWS Glue with a PySpark job
Show Answer & Explanation

Correct Answer: A. AWS Lambda with a Python script

Why A is correct: AWS Lambda is the most cost-effective solution because it operates on a pay-per-execution model with no infrastructure costs when idle. With 1 GB memory requirement and couple-minute execution time per run, Lambda fits perfectly within its limits (up to 10 GB memory, 15 minutes timeout). Lambda natively supports Python, avoiding unnecessary conversion. Since data processing happens once per night per bed, Lambda's pricing model (charging only for actual compute time) is far more economical than continuously running clusters. Why others are wrong: B uses Scala instead of the existing Python script, requiring code conversion C EMR clusters are overkill for this simple processing task and incur costs even when idle D while AWS Glue supports PySpark, it's more expensive than Lambda for infrequent, short-duration jobs

Question 3

A retail company has 15 stores across 6 cities in the United States. Once a month, the sales team requests a visualization in Amazon QuickSight that provides the ability to easily identify revenue trends across cities and stores. The visualization also helps identify outliers that need to be examined with further analysis.Which visual type in QuickSight meets the sales team's requirements?

A. Geospatial chart
B. Line chart
C. Heat map
D. Tree map
Show Answer & Explanation

Correct Answer: C. Heat map

Why C is correct: Heat maps excel at displaying two-dimensional data with color-coded values, making them ideal for visualizing revenue across two categorical dimensions (cities and stores). The color intensity naturally highlights outliers and trends, allowing the sales team to quickly identify high and low performers across the matrix of cities and stores without detailed analysis. Why others are wrong: A Geospatial charts show geographic locations but don't effectively display comparative metrics across categories B Line charts show trends over time, not comparisons across multiple categorical dimensions D Tree maps show hierarchical data and proportions but are less effective at identifying outliers compared to heat maps

Question 4

A gaming company is building a serverless data lake. The company is ingesting streaming data into Amazon Kinesis Data Streams and is writing the data to Amazon S3 through Amazon Kinesis Data Firehose. The company is using 10 MB as the S3 buffer size and is using 90 seconds as the buffer interval. The company runs an AWS Glue ETL job to merge and transform the data to a different format before writing the data back to Amazon S3.Recently, the company has experienced substantial growth in its data volume. The AWS Glue ETL jobs are frequently showing an OutOfMemoryError error.Which solutions will resolve this issue without incurring additional costs? (Choose two.)

A. Place the small files into one S3 folder. Define one single table for the small S3 files in AWS Glue Data Catalog. Rerun the AWS Glue ETL jobs against this AWS Glue table.
B. Create an AWS Lambda function to merge small S3 files and invoke them periodically. Run the AWS Glue ETL jobs after successful completion of the Lambda function.
C. Run the S3DistCp utility in Amazon EMR to merge a large number of small S3 files before running the AWS Glue ETL jobs.
D. Use the groupFiles setting in the AWS Glue ETL job to merge small S3 files and rerun AWS Glue ETL jobs.
E. Update the Kinesis Data Firehose S3 buffer size to 128 MB. Update the buffer interval to 900 seconds
Show Answer & Explanation

Correct Answers: D. Use the groupFiles setting in the AWS Glue ETL job to merge small S3 files and rerun AWS Glue ETL jobs.; E. Update the Kinesis Data Firehose S3 buffer size to 128 MB. Update the buffer interval to 900 seconds

Why D is correct: The groupFiles setting in AWS Glue ETL jobs is specifically designed to address the small files problem by logically grouping multiple small files together during processing. This reduces memory pressure by processing files in batches rather than individually, directly addressing the OutOfMemoryError without additional infrastructure costs. Why E is correct: Increasing the Firehose buffer size to 128 MB and buffer interval to 900 seconds (15 minutes) creates larger files before writing to S3, addressing the root cause of the small files problem. Fewer, larger files reduce the number of objects the Glue job needs to process, preventing memory issues. This is a configuration change with no additional cost. Why others are wrong: A simply organizing files doesn't solve the processing issue B requires Lambda which adds operational overhead and doesn't avoid additional costs (Lambda execution costs) C requires running EMR which incurs significant additional costs

Question 5

A company using Amazon QuickSight Enterprise edition has thousands of dashboards, analyses, and datasets. The company struggles to manage and assign permissions for granting users access to various items within QuickSight. The company wants to make it easier to implement sharing and permissions management.Which solution should the company implement to simplify permissions management?

A. Use QuickSight folders to organize dashboards, analyses, and datasets. Assign individual users permissions to these folders.
B. Use QuickSight folders to organize dashboards, analyses, and datasets. Assign group permissions by using these folders.
C. Use AWS IAM resource-based policies to assign group permissions to QuickSight items.
D. Use QuickSight user management APIs to provision group permissions based on dashboard naming conventions.
Show Answer & Explanation

Correct Answer: B. Use QuickSight folders to organize dashboards, analyses, and datasets. Assign group permissions by using these folders.

Why B is correct: QuickSight folders provide organizational hierarchy for dashboards, analyses, and datasets, and most importantly, support group-level permissions. Assigning permissions to groups (rather than individual users) dramatically simplifies management at scale - you manage permissions once per group rather than thousands of times per user. This is the recommended AWS best practice for large-scale permissions management. Why others are wrong: A using individual user permissions defeats the purpose of simplification when dealing with thousands of users C IAM resource-based policies don't directly control QuickSight item access; QuickSight uses its own permission model D is overly complex and relies on naming conventions rather than using built-in permission features

Ready for the Full DAS-C01 Experience?

Access all 33 pages of practice questions, track your progress, and simulate the real exam with timed mode.

Start Interactive Quiz →