getcertified4sure.com

AWS-Certified-Big-Data-Specialty Exam

Amazon AWS-Certified-Big-Data-Specialty Q&A 2021




Our pass rate is high to 98.9% and the similarity percentage between our AWS-Certified-Big-Data-Specialty study guide and real exam is 90% based on our seven-year educating experience. Do you want achievements in the Amazon AWS-Certified-Big-Data-Specialty exam in just one try? I am currently studying for the Amazon AWS-Certified-Big-Data-Specialty exam. Latest Amazon AWS-Certified-Big-Data-Specialty Test exam practice questions and answers, Try Amazon AWS-Certified-Big-Data-Specialty Brain Dumps First.

Amazon AWS-Certified-Big-Data-Specialty Free Dumps Questions Online, Read and Test Now.

NEW QUESTION 1
Which two AWS services provide out-of-the-box user configurable automatic backup-as-a- service
and backup rotation options? Choose 2 answers

  • A. Amazon S3
  • B. Amazon RDS
  • C. Amazon EBS
  • D. Amazon Redshift

Answer: BD

NEW QUESTION 2
A systems engineer for a company proposes digitalization and backup of large archives for customers.
The systems engineer needs to provide users with a secure storage that makes sure that data will never be tempered with once it has been uploaded. How should this be accomplished?

  • A. Create an Amazon Glacier Vaul
  • B. Specify a “Deny” Vault lock policy on this vault to block “glacier:DeleteArchive”.
  • C. Create an Amazon S3 bucke
  • D. Specify a “Deny” bucket policy on this bucket to block “s3:DeleteObject”.
  • E. Create an Amazon Glacier Vaul
  • F. Specify a “Deny” vault access policy on this Vault to block “glacier:DeleteArchive”.
  • G. Create a secondary AWS containing an Amazon S3 bucke
  • H. Grant “s3:PutObject” to the primary account.

Answer: A

NEW QUESTION 3
A data engineer in a manufacturing company is designing a data processing platform that receives a
large volume of unstructured data. The data engineer must populate a well- structured star schema in Amazon Redshift.
What is the most efficient architecture strategy for this purpose?

  • A. Transform the unstructured data using Amazon EMR and generate CSV dat
  • B. COPY data into the analysis schema within Redshift.
  • C. Load the unstructured data into Redshift, and use string paring functions to extract structured data for inserting into the analysis schema.
  • D. When the data is saved to Amazon S3. Use S3 Event Notifications and AWS Lambda to transform the file conten
  • E. Insert the data into the analysis schema on Redshift.
  • F. Normalize the data using an AWS Marketplace ETL tool persist the result to Amazon S3 and use AWS Lambda to INSERT the data into Redshift.

Answer: B

NEW QUESTION 4
What is one key difference between an Amazon EBS-backed and an instance-store backed instance?

  • A. Amazon EBS-backed instances can be stopped and restarted
  • B. Instance-store backed instances can be stopped and restarted
  • C. Auto scaling requires using Amazon EBS-backed instances
  • D. Virtual Private Cloud requires EBS backed instances

Answer: A

NEW QUESTION 5
You have a load balancer configured for VPC, and all backend Amazon EC2 instances are in service. However, your web browser times out when connecting to the load balancer’s DNS name. Which options are probable causes of this behavior?

  • A. The load balancer was not configured to use a public subnet with an Internet gateway configured
  • B. The Amazon EC2 instances do not have a dynamically allocated private IP address
  • C. The security groups or network ACLs are not properly configured for web traffic
  • D. The load balancer is not configured in a private subnet with a NAT instance
  • E. The VPC does not have a VGW configured

Answer: AC

NEW QUESTION 6
An enterprise customer is migrating to Redshift and is considering using dense storage nodes in its
Redshift cluster. The customer wants to migrate 50 TB of data. The customer’s query patterns involve performing many joins with thousands of rows. The customer needs to know how many nodes are needed in its target Redshift cluster. The customer has a limited budget and needs to avoid performing tests unless absolutely needed. Which approach should this customer use?

  • A. Start with many small nodes
  • B. Start with fewer large nodes
  • C. Have two separate clusters with a mix of small and large nodes
  • D. Insist on performing multiple tests to determine the optimal configuration

Answer: D

NEW QUESTION 7
A user has launched an EC2 instance from an instance store backed AMI. The user has attached an
additional instance store volume to the instance. The user wants to create an AMI from the running instance. Will the AMI have the additional instance store volume data?

  • A. Yes, the block device mapping will have information about the additional instance store volume
  • B. No, since the instance store backed AMI can have only the root volume bundled
  • C. It is not possible to attach an additional instance store volume to the existing instance store backed AMI instance
  • D. No, since this is ephermal storage it will not be a part of the AMI

Answer: A

NEW QUESTION 8
You are configuring your company’s application to use Auto Scaling and need to move user state
information. Which of the following AWS services provides a shared data store with durability and low latency?

  • A. Amazon Simple Storage Service
  • B. Amazon DynamoDB
  • C. Amazon EC2 instance storage
  • D. AWS ElasticCache Memcached

Answer: A

NEW QUESTION 9
Which of the following notification endpoints or clients are supported by Amazon Simple Notification Service? Choose 2 answers

  • A. Email
  • B. CloudFront distribution
  • C. File Transfer Protocol
  • D. Short Message Service
  • E. Simple Network Management Protocol

Answer: BC

NEW QUESTION 10
An organization needs a data store to handle the following data types and access patterns
• Faceting
• Search
• Flexible schema (JSON) and fixed schema
• Noise word elimination
Which data store should the organization choose?

  • A. Amazon Relational Database Service (RDS)
  • B. Amazon Redshift
  • C. Amazon DynamoDB
  • D. Amazon Elasticsearch Service

Answer: C

NEW QUESTION 11
A media advertising company handles a large number of real-time messages sourced from over 200
websites. The company’s data engineer needs to collect and process records in real time for analysis using Spark Streaming on Amazon Elastic MapReduce (EMR). The data engineer needs to fulfill a corporate mandate to keep ALL raw messages as they are received as a top priority.
Which Amazon Kinesis configuration meets these requirements?

  • A. Publish messages to Amazon Kinesis Firehose backed by Amazon Simple Storage Service (S3). Pull messages off Firehose with Spark Streaming in parallel to persistence to Amazon S3
  • B. Publish messages to Amazon Kinesis Stream
  • C. Pull messages off Stream with Spark Streaming in parallel to AWS messages from Streams to Firehose backed by Amazon Simple Storage Service (S3)
  • D. Publish messages to Amazon Kinesis Firehose backed by Amazon Simple Storage (S3). Use AWS Lambda messages from Firehose to Streams for processing with Spark Streaming
  • E. Publish messages to Amazon Kinesis Streams, pull messages off with Spark Streaming and write data new data to Amazon Simple Storage Service (S3) before and after processing

Answer: D

NEW QUESTION 12
A user is trying to setup a recurring Auto Scaling process. The user has setup one process to scale up every day at 8 am and scale down at 7 PM. The user is trying to setup another recurring process which scales up on the 1st of every month at 8 AM and scales down the same day at 7 PM. What will Auto Scaling do in this scenario?

  • A. Auto Scaling will execute both processes but will add just one instance on the 1st
  • B. Auto Scaling will add two instances on the 1st of the month
  • C. Auto Scaling will schedule both the processes but execute only one process randomly
  • D. Auto Scaling will throw an error since there is a conflict in the schedule of two separate Auto Scaling Processes

Answer: D

NEW QUESTION 13
Managers in a company need access to the human resources database that runs on Amazon Redshift, to run reports about their employees. Managers must only see information about their direct reports.
Which technique should be used to address this requirement with Amazon Redshift?

  • A. Define an IAM group for each employee as an IAM user in that group and use that to limit theaccess.
  • B. Use Amazon Redshift snapshot to create one cluster per manage
  • C. Allow the managers to access only their designated clusters.
  • D. Define a key for each manager in AWS KMS and encrypt the data for their employees with their private keys.
  • E. Define a view that uses the employee’s manager name to filter the records based on current user names.

Answer: B

NEW QUESTION 14
A company that provides economics data dashboards needs to be able to develop software to display
rich, interactive, data-driven graphics that run in web browsers and leverages the full stack of web standards (HTML, SVG and CSS).
Which technology provides the most appropriate for this requirement?

  • A. D3.js
  • B. Python/Jupyter
  • C. R Studio
  • D. Hue

Answer: C

NEW QUESTION 15
You have an ASP.NET web application running in Amazon Elastic BeanStalk. Your next version of the
application requires a third-party Windows installer package to be installed on the instance on first boot and before the application launches.
Which options are possible? Choose 2 answer

  • A. In the application’s Global.asax file, run msiexec.exe to install the package using Process.Start() in the Application_Start event handler
  • B. In the source bundle’s .ebextensions folder, create a file with a .config extensio
  • C. In the file, under the “packages” section and “msi” package manager, include the package’s URL
  • D. Launch a new Amazon EC2 instance from the AMI used by the environmen
  • E. Log into the instance, install the package and run syspre
  • F. Create a new AM
  • G. Configure the environment to use the new AMI
  • H. In the environment’s configuration, edit the instances configuration and add the package’s URL to the “Packages” section
  • I. In the source bundle’s .ebextensions folder, create a “Packages” folde
  • J. Place the package in the folder

Answer: BC

NEW QUESTION 16
A company receives data sets coming from external providers on Amazon S3. Data sets from different providers are dependent on one another. Data sets will drive at different and is no particular order.
A data architect needs to design a solution that enables the company to do the following:
• Rapidly perform cross data set analysis as soon as the data becomes available
• Manage dependencies between data sets that arrives at different times
Which architecture strategy offers a scalable and cost-effective solution that meets these requirements?

  • A. Maintain data dependency information in Amazon RDS for MySQ
  • B. Use an AWS Pipeline job to load an Amazon EMR Hive Table based on task dependencies and event notification triggers in Amazon S3
  • C. Maintain data dependency information in an Amazon DynamoDB tabl
  • D. Use Amazon SNS and event notification to publish data to a fleet of Amazon EC2 worker
  • E. Once the task dependencies have been resolved process the data with Amazon EMR
  • F. Maintain data dependency information in an Amazon ElasticCache Redis cluste
  • G. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to Redi
  • H. Once the dependencies have been resolved process the data with Amazon EMR
  • I. Maintain data dependency information in an Amazon DynamoDB tabl
  • J. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to the task associated with it in DynamoD
  • K. Once all task dependencies have been resolved process the data with Amazon EMR

Answer: D

NEW QUESTION 17
You have been asked to use your department’s existing continuous integration (CI) tool to test a
three- tier web architecture defined in an AWS CloudFormation template. The tool already supports AWS APIs and can launch new AWS CloudFormation stacks after polling version control. The CI tool reports on the success of the AWS CloudFormation stack creation by using the DescribeStacks API to look for the CREATE_COMPLETE status.
The architecture tiers defined in the template consist of:
. One load balancer
. Five Amazon EC2 instances running the web application
. One multi-AZ Amazon RDS instance How would you implement this? Choose 2 answers

  • A. Define a WaitCondition and a WaitConditionhandle for the output of a output of a UserData command that does sanity checking of the application’s post-install state
  • B. Define a CustomResource and write a script that runs architecture-level integration tests through the load balancer to the application and database for the state of multiple tiers
  • C. Define a WaitCondition and use a WaitConditionHandle that leverages the AWS SDK to run the DescribeStacks API call until the CREATE_COMPLETE status is returned
  • D. Define a CustomResource that leverages the AWS SDK to run the DescribeStacks API call until the CREATE_COMPLETE status is returned
  • E. Define a UserDataHandle for the output of a UserData command that does sanity checking of the application’s post-install state and runs integration tests on the state of multiple tiers through load balancer to the application
  • F. Define a UserDataHandle for the output of a CustomResource that does sanity checking of the application’s post-install state

Answer: AF

NEW QUESTION 18
You run a small online consignment marketplace. Interested sellers complete an online application in
order to allow them to sell their products on your website. Once approved, they can their product using a custom interface. From that point, you manage the shopping cart process so that when a buyer decides to buy a product, you handle the billing and coordination the shipping. Part of this process requires sending emails to the buyer and the seller at different stages. Your system has been running on AWS for a few months. Occasionally, products are shipped before payment has cleared and emails are sent out of order. Furthermore, sometimes credit cards are being charged twice.
How can you resolve these problems?

  • A. Use the Amazon Simple Queue Service (SQS), and use a different set of workers for each task
  • B. Use the Amazon Simple Workflow Service (SWF), and use a different set of workers for each task.
  • C. Use the Simple Email Service (SES) to control the correct order of email delivery
  • D. Use the AWS Data Pipeline service to control the process flow of the various tasks
  • E. Use the Amazon Simple Queue Service (SQS), and use a single set of workers for each task

Answer: E

NEW QUESTION 19
You have been asked to handle a large data migration from multiple Amazon RDS MySQL instances to
a DynamoDB table. You have been given a short amount of time to complete the data migration. What will allow you to complete this complex data processing workflow?

  • A. Create an Amazon Kinesis data stream, pipe in all of the Amazon RDS data, and direct data toward DynamoDB table
  • B. Write a script in you language of choice, install the script on an Amazon EC2 instance, and then use Auto Scaling groups to ensure that the latency of the mitigation pipelines never exceeds four seconds in any 15-minutes period.
  • C. Write a bash script to run on your Amazon RDS instance that will export data into DynamoDB
  • D. Create a data pipeline to export Amazon RDS data and import the data into DynamoDB

Answer: D

NEW QUESTION 20
You need to configure an Amazon S3 bucket to serve static assets for your public-facing web application. Which methods ensure that all objects uploaded to the bucket are set to public read? Choose 2 answers

  • A. Set permissions on the object to public read during upload
  • B. Configure the bucket ACL to sell all objects to public read
  • C. Configure the bucket policy to set all objects to public read
  • D. Use AWS identity and access Management roles to set the bucket to public read
  • E. Amazon S3 objects default to public read, so no action is needed

Answer: BC

NEW QUESTION 21
An administrator receives about 100 files per hour into Amazon S3 and will be loading the files into Amazon Redshift. Customers who analyze the data within Redshift gain significant value when they receive data as quickly as possible. The customers have agreed to a maximum loading interval of 5 minutes. Which loading approach should the administrator use to meet this objective?

  • A. Load each file as it arrives because getting data into the cluster as quickly as possible is the priority.
  • B. Load the cluster as soon as the administrator has the same number of files as nodes in the cluster.
  • C. Load the cluster when the administrator has an even multiple of files relative to Cluster Slice Count, or 5 minutes whichever comes first.
  • D. Load the cluster when the number files is less than the Cluster Slice Count.

Answer: C

NEW QUESTION 22
A user has created a launch configuration for Auto Scaling where CloudWatch detailed monitoring is
disabled. The user wants to now enable detailed monitoring. How can the user achieve this?

  • A. Update the Launch config with CLI to set InstanceMonitoringDisabled = false
  • B. The user should change the Auto Scaling group from the AWS console to enable detailed monitoring
  • C. Update the Launch config with CLI to set InstanceMonitoring.Enabled = true
  • D. Create a new Launch Config with detail monitoring enabled and update the Auto Scaling group

Answer: D

NEW QUESTION 23
Using only AWS services. You intend to automatically scale a fleet of stateless of stateless web
servers based on CPU and network utilization metrics. Which of the following services are needed? Choose 2 answers

  • A. Auto Scaling
  • B. Amazon Simple Notification Service
  • C. AWS Cloud Formation
  • D. CloudWatch
  • E. Amazon Simple Workflow Service

Answer: AD

NEW QUESTION 24
You have a video Trans coding application running on Amazon EC2. Each instance pools a queue to find out which video should be Trans coded, and then runs a Trans coding process.
If this process is interrupted, the video will be Trans coded by another instance based on the queuing system. You have a large backlog of videos which need to be Trans coded and would like to reduce this backlog by adding more instances. You will need these instances only until the backlog is reduced. Which type of Amazon EC2 instance should you use to reduce the backlog in the most cost- effective way?

  • A. Dedicated instances
  • B. Spot instances
  • C. On-demand instances
  • D. Reserved instances

Answer: B

NEW QUESTION 25
An administrator needs to design a distribution strategy for a star schema in a Redshift cluster. The
administrator needs to determine the optimal distribution style for the tables in the Redshift schema. In which three circumstances would choosing Key-based distribution be most appropriate? (Select three)

  • A. When the administrator needs to optimize a large, slowly changing dimension table
  • B. When the administrator needs to reduce cross-node traffic
  • C. When the administrator needs to optimize the fact table for parity with the number of slices
  • D. When the administrator needs to balance data distribution and collocation of data
  • E. When the administrator needs to take advantage of data locality on a local node of joins and aggregates

Answer: ADE

NEW QUESTION 26
......

P.S. Easily pass AWS-Certified-Big-Data-Specialty Exam with 243 Q&As Certifytools Dumps & pdf Version, Welcome to Download the Newest Certifytools AWS-Certified-Big-Data-Specialty Dumps: https://www.certifytools.com/AWS-Certified-Big-Data-Specialty-exam.html (243 New Questions)