Browse DBS Questions
Study all 100 questions at your own pace with detailed explanations
Total: 100 questionsPage: 5 of 10
Question 41 of 100
An administrator is processing events in near real-time using Kinesis streams and Lambda. Lambda intermittently fails to process batches from one of the shards due to a 15-minute time limit. What is a possible solution for this problem?
AAdd more Lambda functions to improve concurrent batch processing.
BReduce the batch size that Lambda is reading from the stream.
CIgnore and skip events that are older than 15 minutes and put them to Dead Letter Queue (DLQ).
DConfigure Lambda to read from fewer shards in parallel.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 42 of 100
A company is using Kinesis data streams to store the log data, which is processed by an application every 12 hours. As the data needs to reside in Kinesis data streams for 12 hours, the Security team wants the data to be encrypted at rest. How can it be secured in a most efficient way?
AKinesis does not support encryption
BEncrypt using SSL/TLS for encrypting the data.
CEncrypt using S3 Server Side Encryption.
DEncrypt using Kinesis Server Side Encryption.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 43 of 100
A company needs a churn prevention model to predict which customers will NOT renew their yearly subscription to the company’s service. The company plans to provide these customers with a promotional offer. A binary classification model that uses Amazon Machine Learning is required. On which basis should this binary classification model be built?
AUser profiles (age, gender, income, occupation)
BLast user session
CEach user time series events in the past 3 months
DQuarterly results
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 44 of 100
A company launched EMR cluster to support their big data analytics requirements. They have multiple data sources built out of S3, SQL databases, MongoDB, Redis, RDS, other file systems. They are looking for distributed processing framework and programming model that helps you do machine learning, stream processing, or graph analytics using Amazon EMR clusters Which EMR Hadoop ecosystem fulfils the requirements?
AApache Hive
BApache HBase
CApache HCatalog
DApache Spark
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 45 of 100
Your company produces customer commissioned one-of-a-kind skiing helmets combining high fashion with custom technical enhancements. Customers can show off their Individuality on the ski slopes and have access to head-up-displays. GPS rear-view cams and any other technical innovation they wish to embed in the helmet. The current manufacturing process is data rich and complex including assessments to ensure that the custom electronics and materials used to assemble the helmets are to the highest standards. Assessments are a mixture of human and automated assessments you need to add a new set of assessment to model the failure modes of the custom electronics using GPUs with CUDA across a cluster of servers with low latency networking. What architecture would allow you to automate the existing process using a hybrid approach and ensure that the architecture can support the evolution of processes over time?
AUse AWS Data Pipeline to manage movement of data & meta-data and assessments Use an auto-scaling group of G2 instances in a placement group.
BUse Amazon Simple Workflow (SWF) to manage assessments, movement of data & meta-data. Use an autoscaling group of G2 instances in a placement group.
CUse Amazon Simple Workflow (SWF) to manage assessments, movement of data & meta-data. Use an autoscaling group of C3 instances with SR-IOV (Single Root I/O Virtualization).
DUse AWS data Pipeline to manage movement of data & meta-data and assessments use auto-scaling group of C3 with SR-IOV (Single Root I/O virtualization)
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 46 of 100
A company operates an international business served from a single AWS region. The company wants to expand into a new country. The regulator for that country requires the Data Architect to maintain a log of financial transactions in the country within 24 hours of the product transaction. The production application is latency insensitive. The new country contains another AWS region. What is the most cost-effective way to meet this requirement?
AUse CloudFormation to replicate the production application to the new region.
BUse Amazon CloudFront to serve application content locally in the country; Amazon CloudFront logs will satisfy the requirement.
CContinue to serve customers from the existing region while using Amazon Kinesis to stream transaction data to the regulator.
DUse Amazon S3 cross-region replication to copy and persist production transaction logs to a bucket in the new country’s region.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 47 of 100
You have recently joined a startup company building sensors to measure street noise and air quality in urban areas. The company has been running a pilot deployment of around 100 sensors for 3 months. Each sensor uploads 1KB of sensor data every minute to a backend hosted on AWS. During the pilot, you measured a peak or 10 IOPS on the database, and you stored an average of 3GB of sensor data per month in the database. The current deployment consists of a load-balanced auto scaled Ingestion layer using EC2 instances and a PostgreSQL RDS database with 500GB standard storage. The pilot is considered a success and your CEO has managed to get the attention or some potential investors. The business plan requires a deployment of at least 100K sensors, which needs to be supported by the backend. You also need to store sensor data for at least two years to be able to compare year over year Improvements. To secure funding, you have to make sure that the platform meets these requirements and leaves room for further scaling. Which setup will meet the requirements?
AAdd an SQS queue to the ingestion layer to buffer writes to the RDS instance
BIngest data into a DynamoDB table and move old data to a Redshift cluster
CReplace the RDS instance with a 6 node Redshift cluster with 96TB of storage
DKeep the current architecture but upgrade RDS storage to 3TB and 10K provisioned IOPS
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 48 of 100
A company receives data sets coming from external providers on Amazon S3. Data sets from different providers are dependent on one another. Data sets will arrive at different times and in no particular order. A data architect needs to design a solution that enables the company to do the following: Rapidly perform cross data set analysis as soon as the data become available Manage dependencies between data sets that arrive at different times Which architecture strategy offers a scalable and cost-effective solution that meets these Requirements?
AMaintain data dependency information in Amazon RDS for MySQL. Use an AWS Data Pipeline job to load an Amazon EMR Hive table based on task dependencies and event notification triggers in Amazon S3.
BMaintain data dependency information in an Amazon DynamoDB table. Use Amazon SNS and event notifications to publish data to fleet of Amazon EC2 workers. Once the task dependencies have been resolved, process the data with Amazon EMR.
CMaintain data dependency information in an Amazon ElastiCache Redis cluster. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to Redis. Once the task dependencies have been resolved, process the data with Amazon EMR.
DMaintain data dependency information in an Amazon DynamoDB table. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to the task associated with it in DynamoDB. Once all task dependencies have been resolved, process the data with Amazon EMR.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 49 of 100
A media advertising company handles a large number of real-time messages sourced from over 200 websites in real time. Processing latency must be kept low. Based on calculations, a 60-shard Amazon Kinesis stream is more than sufficient to handle the maximum data throughput, even with traffic spikes. The company also uses an Amazon Kinesis Client Library (KCL) application running on Amazon Elastic Compute Cloud (EC2) managed by an Auto Scaling group. Amazon CloudWatch indicates an average of 25% CPU and a modest level of network traffic across all running servers. The company reports a 150% to 200% increase in latency of processing messages from Amazon Kinesis during peak times. There are NO reports of delay from the sites publishing to Amazon Kinesis. What is the appropriate solution to address the latency?
AIncrease the number of shards in the Amazon Kinesis stream to 80 for greater concurrency.
BIncrease the size of the Amazon EC2 instances to increase network throughput.
CIncrease the minimum number of instances in the Auto Scaling group.
DIncrease Amazon DynamoDB throughput on the checkpoint table.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation
Question 50 of 100Multiple Choice
An administrator needs to design a strategy for the schema in a Redshift cluster. The administrator needs to determine the optimal distribution style for the tables in the Redshift schema. In which two circumstances would choosing EVEN distribution be most appropriate? (Choose two.)
AWhen the tables are highly denormalized and do NOT participate in frequent joins.
BWhen data must be grouped based on a specific key on a defined slice.
CWhen data transfer between nodes must be eliminated.
DWhen a new table has been loaded and it is unclear how it will be joined to dimension.
💡 Try to answer first, then click "Show Answer" to see the correct answer and explanation