Want to know Exambible 70-470 Exam practice test features? Want to lear more about Microsoft Recertification for MCSE: Business Intelligence certification experience? Study Precise Microsoft 70-470 answers to Renovate 70-470 questions at Exambible. Gat a success with an absolute guarantee to pass Microsoft 70-470 (Recertification for MCSE: Business Intelligence) test on your first attempt.
Q171. - (Topic 10)
You are developing a SQL Server Analysis Services (SSAS) cube.
The data warehouse has a table named FactStock that is used to track movements of stock. A column named MovementQuantity contains quantities of stock. A positive quantity is used for input and negative quantity is used for output. A column named MovementDate is related to the time dimension. The quantity in stock, at a given point in time, can be evaluated as the sum of all MovementQuantity values at that point in time.
You need to create a measure that calculates the quantity in stock value.
What should you do?
A. Use role playing dimensions.
B. Use the Business Intelligence Wizard to define dimension intelligence.
C. Add a measure that uses the Count aggregate function to an existing measure group.
D. Add a measure that uses the DistinctCount aggregate function to an existing measure group.
E. Add a measure that uses the LastNonEmpty aggregate function. Use a regular relationship between the time dimension and the measure group.
F. Add a measure group that has one measure that uses the DistinctCount aggregate function.
G. Add a calculated measure based on an expression that counts members filtered by the Exists and NonEmpty functions.
H. Add a hidden measure that uses the Sum aggregate function. Add a calculated measure aggregating the measure along the time dimension.
I. Create several dimensions. Add each dimension to the cube.
J. Create a dimension. Then add a cube dimension and link it several times to the measure group.
K. Create a dimension. Create regular relationships between the cube dimension and the measure group. Configure the relationships to use different dimension attributes.
L. Create a dimension with one attribute hierarchy. Set the IsAggregatable property to False and then set the DefaultMember property. Use a regular relationship between the dimension and measure group.
M. Create a dimension with one attribute hierarchy. Set the IsAggregatable property to False and then set the DefaultMember property. Use a many-to-many relationship to link the dimension to the measure group.
N. Create a dimension with one attribute hierarchy. Set the ValueColumn property, set the IsAggregatable property to False, and then set the DefaultMember property. Configure the cube dimension so that it does not have a relationship with the measure group. Add a calculated measure that uses the MemberValue attribute property.
O. Create a new named calculation in the data source view to calculate a rolling sum. Add a measure that uses the Max aggregate function based on the named calculation.
Answer: H
Q172. HOTSPOT - (Topic 9)
You are developing a SQL Server Reporting Services (SSRS) solution.
You plan to create reports based on the results of a currency exchange SOAP web service
call.
You need to configure a shared data source.
Which data source type should you use?
To answer, select the appropriate type from the drop-down list in the answer area.
Answer:
Q173. - (Topic 9)
A company runs SQL Server Database Engine and SQL Server Reporting Services (SSRS) in native mode. Reports are based on data that is cached in multiple shared datasets. Source data is archived each day at midnight for regulatory compliance purposes. The shared datasets may continue to cache data that should not be used in reports. Shared report schedules are often paused during nightly server maintenance windows.
Reports must not return archived data.
You need to create a fully automated solution to ensure that reports do not deliver archived data.
What should you do? (More than one answer choice may achieve the goal. Select the BEST answer.)
A. Write a script that calls the flushcache method to clear individual items from the SSRS cache. Create a SQL Server Agent job that runs rs.exe with the script as an input file, and schedule the job to run every day after the archive process completes.
B. Create a SQL Server Agent job that uses a Transact-SQL (T-SQL) step to truncate the dbo.ExecutionCache table in the ReportServerTempDB database. Schedule the job to run every day after the archive process completes.
C. Create a SQL Server Agent job that restarts the SQL Server Reporting Services service. Schedule the job to run every day after the archive process completes.
D. Create a shared schedule. Configure the datasets to expire on the shared schedule.
Answer: A
Q174. - (Topic 10)
You are conducting a design review of a multidimensional project.
In the geography dimension, all non-key attributes relate directly to the key attribute. The
underlying data of the geography dimension supports relationships between attributes.
You need to increase query and dimension processing performance.
What should you do?
A. For the geography dimension, set the ProcessingMode property to LazyAggregations
B. For the dimension attributes of the geography dimension, define appropriate attribute relationships.
C. For the geography dimension, set the ProcessingPriority property to 1.
D. For the dimension attributes of the geography dimension, set the GroupingBehavior property to EncourageGrouping.
Answer: B
Q175. DRAG DROP - (Topic 9)
You are creating a SQL Server Integration Services (SSIS) package to populate a fact table from a source table. The fact table and source table are located in a Windows Azure SQL Database. The source table has a price field and a tax field. The OLE DB source uses the data access mode of Table.
You have the following requirements:
. The fact table must populate a column named TotalCost that computes the sum of the price and tax columns. . Before the sum is calculated, any records that have a price of zero must be discarded.
You need to create the SSIS package in SQL Server Data Tools.
In what sequence should you order four of the listed components for the data flow task? (To answer, move the appropriate components from the list of components to the answer area and arrange them in the correct order.)
Answer:
Q176. DRAG DROP - (Topic 9)
You develop a SQL Server Integration Services (SSIS) project named Project1 by using
SQL Server Data Tools (SSDT). Project1 contains a package named Package1.
You add a project parameter named EnvironmentText to Project1.
In SQL Server Management Studio (SSMS), you create the SSIS catalog.
You have the following requirements:
. Configure Package1 so it can run within either a development environment or a test environment.
. Pass the value of an SSIS environment variable to the EnvironmentText project parameter.
. The value of the environment variable must be different for each SSIS environment.
You need to deploy the SSIS project and configure the SSIS environment.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Q177. - (Topic 8)
You need to recommend a cube architecture for CUBE1. The solution must meet the performance requirements for CUBE1.
Which two partitions should you recommend creating? Each correct answer presents part of the solution.
A. Partitions based on the values of the customerID column in the dimension table
B. Partitions based on the values of the customerID column in the fact table
C. Partitions based on the values of the productID column in the fact table
D. Partitions based on the values of the productID column in the dimension table
Answer: A,D
Topic 9, Mix Questions Set A
Q178. - (Topic 10)
You are developing a SQL Server Analysis Services (SSAS) cube for the sales department at your company.
The sales department requires the following set of metrics:
Unique count of customers
Unique count of products sold
Sum of sales
You need to ensure that the cube meets the requirements while optimizing query response time.
What should you do? (Each answer presents a complete solution. Choose all that apply.)
A. Use the Define Semiadditive Behavior page to enable semi-additive behavior.
B. Use ROLAP storage for all partitions.
C. Use the Define Semiadditive Behavior page to disable semi-additive behavior.
D. Use the Distinct Count and Sum measure aggregation functions.
E. Place the distinct count measures in separate measure groups.
F. Create a single measure group containing all measures
Answer: A,E
Explanation: A: Semiadditive Function Select the aggregation function for the selected measure. The aggregate functions available include DistinctCount, Aggregated using the DistinctCount function. E (not D, Not F): Typically, the best performance occurs when each distinct count measure is in its own measure group, and that measure group has the same dimensionality as the initial measure group.
Q179. - (Topic 6)
You need to design the dimCustomers table.
Which design approach should you use?
A. Reference dimension
B. Type 2 slowly changing dimension
C. Junk dimension
D. Conformed dimension
E. Type 1 slowly changing dimension
Answer: B Topic 7, Contoso, Ltd Case B
General Background You are the business intelligence (BI) solutions architect for Contoso, Ltd, an online retailer.
You produce solutions by using SQL Server 2012 Business Intelligence edition and Microsoft SharePoint Server 2010 Service Pack 1 (SP1) Enterprise edition.
A SharePoint farm has been installed and configured for intranet access only. An Internet-facing web server hosts the company's public e-commerce website. Anonymous access is not configured on the Internet-facing web server.
Data Warehouse The data warehouse is deployed on a 5QL Server 2012 relational database instance. The data warehouse is structured as shown in the following diagram.
The following Transact-SQL (T-SQL) script is used to create the FactSales and FactPopulation tables:
The FactPopulation table is loaded each year with data from a Windows Azure Marketplace commercial dataset. The table contains a snapshot of the population values for all countries of the world for each year. The world population for the last year loaded exceeds
6.8 billion people.
ETL Process SQL Server Integration Services (SSIS) is used to load data into the data warehouse. All SSIS projects are developed by using the project deployment model.
A package named StageFactSales loads data into a data warehouse staging table. The package sources its data from numerous CSV files exported from a mainframe system. The CSV file names begin with the letters GLSD followed by a unique numeric identifier that never exceeds six digits. The data content of each CSV file is identically formatted.
A package named LoadFactFreightCosts sources data from a Windows Azure SQL Database database that has data integrity problems. The package may retrieve duplicate rows from the database.
The package variables of all packages have the RaiseChangedEvent property set to true. A package-level event handler for the OnVariableValueChanged event consists of an Execute SQL task that logs the System::VariableName and System::VariableValue variables.
Data Models SQL Server Analysis Services (SSAS) is used to host the Corporate BI multidimensional database. The Corporate BI database contains a single data source view named Data Warehouse. The Data Warehouse data source view consists of all data warehouse tables. All data source view tables have been converted to named queries.
The Corporate BI database contains a single cube named Sales Analysis and three database dimensions: Date, Customer and Product. The dimension usage for the Sales Analysis cube is as shown in the following image.
The Customer dimension contains a single multi-level hierarchy named Geography. The structure of the Geography hierarchy is shown in the following image.
The Sales Analysis cube's calculation script defines one calculated measure named Sales Per Capita. The calculated measure expression divides the Revenue measure by the Population measure and multiplies the result by 1,000. This calculation represents revenue per 1,000 people.
The Sales Analysis cube produces correct Sales Per Capita results for each country of the world; however, the Grand Total for all countries is incorrect, as shown in the following image (rows 2-239 have been hidden).
A role named Analysts grants Read permission for the Sales Analysis cube to all sales and marketing analysts in the company.
SQL Server Reporting Services (SSRS) is configured in SharePoint integrated mode. All reports are based on shared data sources.
Corporate logo images used in reports were originally configured as data-bound images sourced from a SQL Server relational database table. The image data has been exported to JPG files. The image files are hosted on the Internet-facing web server. All reports have been modified to reference the corporate logo images by using the fully qualified URLs of the image files. A red X currently appears in place of the corporate logo in reports.
Users configure data alerts on certain reports. Users can view a report named Sales Profitability on demand; however, notification email messages are no longer being sent when Sales Profitability report data satisfies alert definition rules. The alert schedule settings for the Sales Profitability report are configured as shown in the following image.
Business Requirements Data Models
Users must be able to: . Provide context to measures and filter measures by using all related data warehouse dimensions. . Analyze measures by order date or ship date.
Additionally, users must be able to add a measure named Sales to the report canvas by clicking only once in the Power View field list. The Sales measure must allow users to analyze the sum of the values in the Revenue column of the FactSales data warehouse table. Users must be able to change the aggregation function of the Sales measure.
Analysis and Reporting A sales manager has requested the following query results from the Sales Analysis cube for the 2012 fiscal year:
. Australian postal codes and sales in descending order of sales. . Australian states and the ratio of sales achieved by the 10 highest customer sales made for each city in that state.
Technical Requirements ETL Processes If an SSIS package variable value changes, the package must log the variable name and the new variable value to a custom log table.
The StageFactSales package must load the contents of all files that match the file name pattern. The source file name must also be stored in a column of the data warehouse staging table. In the design of the LoadFactSales package, if a lookup of the dimension surrogate key value for the product code fails, the row details must be emailed to the data steward and written as an error message to the SSIS catalog log by using the public API.
You must configure the LoadFactFreightCosts package to remove duplicate rows, by using the least development effort.
Data Models Users of the Sales Analysis cube frequently filter on the current month's data. You must ensure that queries to the Sales Analysis cube default to the current month in the Order Date dimension for all users.
You must develop and deploy a tabular project for the exclusive use as a Power View reporting data source. The model must be based on the data warehouse. Model table names must exclude the Dim or Fact prefixes. All measures in the model must format values to display zero decimal places.
Analysis and Reporting Reports must be developed that combine the SSIS catalog log messages with the package variable value changes.