2. Log on to www.testpreptraining.com for real exam dumps
AWS – Certified Big Data Specialty Exam
Dumps
1. A web application emits multiple types of events to Amazon Kinesis
Streams for operational reporting. Critical events must be captured
immediately before processing can continue, but informational events
do not need to delay processing. What is the most appropriate solution
to record these different types of events?
A. Log all events using the Kinesis Producer Library
B. Log critical events using the Kinesis Producer Library, and log informational events
using the PutRecords API method.
C. Log critical events using the PutRecords API method, and log informational events
using the Kinesis Producer Library.
D. Log all events using the PutRecords API method.
2. A data engineer needs to collect data from multiple Amazon Redshift
clusters within a business and consolidate the data into a single central
data warehouse. Data must be encrypted at all times while at rest or in
flight. What is the most scalable way to build this data collection
process?
A. Run an ETL process that connects to the source clusters using SSL to issue a SELECT
query for new data, and then write to the target data warehouse using an INSERT
command over another SSL secured connection.
B. Use AWS KMS data key to run an UNLOAD ENCRYPTED command that stores the
data in an unencrypted S3 bucket; run a COPY command to move the data into the target
cluster.
C. Run an UNLOAD command that stores the data in an S3 bucket encrypted with an
AWS KMS data key; run a COPY command to move the data into the target cluster.
D. Connect to the source cluster over an SSL client connection, and write data records to
Amazon Kinesis Firehose to load into your target data warehouse.
3. A data engineer in a manufacturing company is designing a data
processing platform that receives a large volume of unstructured data.
The data engineer must populate a well-structured star schema in
Amazon Redshift. What is the most efficient architecture strategy for
this purpose?
A. Transform the unstructured data using Amazon EMR and generate CSV data. COPY
the CSV data into the analysis schema within Redshift.
3. Log on to www.testpreptraining.com for real exam dumps
B. Load the unstructured data into Redshift, and use string parsing functions to extract
structured data for inserting into the analysis schema.
C. When the data is saved to Amazon S3, use S3 Event Notifications and AWS Lambda to
transform the file contents. Insert the data into the analysis schema on Redshift.
D. Normalize the data using an AWS Marketplace ETL tool, persist the results to Amazon
S3, and use AWS Lambda to INSERT the data into Redshift.
4. A Redshift data warehouse has different user teams that need to
query the same table with very different query types. These user teams
are experiencing poor performance. Which action improves
performance for the user teams in this situation?
A. Create custom table views.
B. Add interleaved sort keys per team.
C. Maintain team-specific copies of the table.
D. Add support for workload management queue hopping.
5. A travel website needs to present a graphical quantitative summary of
its daily bookings to website visitors for marketing purposes. The
website has millions of visitors per day, but wants to control costs by
implementing the least-expensive solution for this visualization. What is
the most cost-effective solution?
A. Generate a static graph with a transient EMR cluster daily, and store it an Amazon S3.
B. Generate a graph using MicroStrategy backed by a transient EMR cluster.
C. Implement a Jupyter front-end provided by a continuously running EMR cluster
leveraging spot instances for task nodes.
D. Implement a Zeppelin application that runs on a long-running EMR cluster.
6. Your application generates a 1 KB JSON payload that needs to be
queued and delivered to EC2 instances for applications. At the end of the
day, the application needs to replay the data for the past 24 hours. In the
near future, you also need the ability for other multiple EC2 applications
to consume the same stream concurrently. What is the best solution for
this?
A. Kinesis Data Streams
B. Kinesis Firehose
C. SNS
D. SQS
4. Log on to www.testpreptraining.com for real exam dumps
7. You are deploying an application to track GPS coordinates of delivery
trucks in the United States.Coordinates are transmitted from each
delivery truck once every three seconds. You need to design an
architecture that will enable real-time processing of these coordinates
from multiple consumers. Which service should you use to implement
data ingestion?
A. Amazon Kinesis
B. AWS Data Pipeline
C. Amazon AppStream
D. Amazon Simple Queue Service
8. You need to provide customers with rich visualizations that allow you
to easily connect multiple disparate data sources in S3, Redshift, and
several CSV files. Which tool should you use that requires the least
setup?
A. Hue on EMR
B. Redshift
C. QuickSight
D. Elasticsearch
9. You've been asked by the VP of People to showcase the current
breakdown of the headcount for each department within your
organization. What chart do you select to do this to make it easy to
compare each department?
A. Line chart
B. Column chart
C. Pie chart
D. Scatter plot
10. Management has requested a comparison of total sales performance
in the five North American regions in January. They're hoping to
determine how to allocate a budget to regions based on performance in
that single period. What sort of visualization do you use in Amazon
QuickSight?
A. Bar chart
B. Line chart
C. Stacked area chart
D. Histogram
5. Log on to www.testpreptraining.com for real exam dumps
Answers
1. (C) Log critical events using the PutRecords API method, and log informational
events using the Kinesis Producer Library.
2. (B) Use AWS KMS data key to run an UNLOAD ENCRYPTED command that
stores the data in an unencrypted S3 bucket; run a COPY command to move the
data into the target cluster.
3. (A) Transform the unstructured data using Amazon EMR and generate CSV data.
COPY the CSV data into the analysis schema within Redshift.
4. (B) Add interleaved sort keys per team.
5. (A) Generate a static graph with a transient EMR cluster daily, and store it an
Amazon S3.
6. (A) Kinesis Data Streams
7. (A) Amazon Kinesis
8. (C) QuickSight
9. (C) Pie chart
10. (A) Bar chart