When you are planning to get your Microsoft Microsoft certification but never know how to prepare for it, Actualtests can offer an individual an absolute best and quick way. We now have a mass associated with Microsoft 70-463 exam brain dumps. For our Microsoft 70-463 instruction course, youll get every one of the 70-463 important contents. Actualtestss Microsoft 70-463 exam questions as well as answers are put together by thousands associated with successful certified professionals around the globe. You will obtain a full spectrum in the Microsoft 70-463 actual test. Master all the key topics of the Microsoft preparation materials, you are ensured for you to pass the genuine exam.
Q71. You are reviewing the design of an existing fact table named factSales, which is loaded incrementally from a SQL Azure database by a SQL Server Integration Services (SSIS) package each hour. The fact table has approximately 4 billion rows and is dimensioned by product, sales date, and sales time of day.
The database administrator is concerned about the rapid growth of the database and users experience poor reporting performance against this database. Reporting requirements have recently changed and the only remaining report that uses this fact table reports sales by product name, sale month, and sale year. No other reports will be created against this table.
You need to reduce the report processing time and minimize the growth of the database.
What should you do?
A. Create an indexed view over the fact table to sum orderTotal by month.
B. Create a view over the fact table to sum orderTotal by month.
C. Change the granularity of the fact table to month.
D. Partition the fact table by productKey.
Answer: B
Q72. You develop a SQL Server Integration Services (SSIS) package that imports Windows Azure SQL Database data into a data warehouse every night.
The Windows Azure SQL Database data contains many misspellings and variations of abbreviations. To import the data, a developer used the Fuzzy Lookup transformation to choose the closest-matching string from a reference table of allowed values. The number of rows in the reference table is very large.
If no acceptable match is found, the Fuzzy Lookup transformation passes a null value.
The current setting for the Fuzzy Lookup similarity threshold is 0.50.
Many values are incorrectly matched.
You need to ensure that more accurate matches are made by the Fuzzy Lookup transformation without degrading performance.
What should you do?
A. Change the similarity threshold to 0.40.
B. Decrease the maximum number of matches per lookup.
C. Change the similarity threshold to 0.85.
D. Increase the maximum number of matches per lookup.
Answer: C
Explanation: * Similarity threshold Set the similarity threshold at the component level by using the slider. The closer the value is to 1, the closer the resemblance of the lookup value to the source value must be to qualify as a match. Increasing the threshold can improve the speed of matching since fewer candidate records need to be considered.
Incorrect:
* Maximum number of matches to output per lookup
Specify the maximum number of matches the transformation can return for each input row.
The default is 1.
Q73. You are developing a SQL Server Integration Services (SSIS) package.
The package is stored as the myPackage.dtsx file in the root directory of the C drive of the
SSIS server.
You need to run the package from the command prompt.
Which command should you use?
A. dtexec /sql myPackage
B. dtexec /d "C:\File System\myPackage.dtsx"
C. dtexec /f "C:\myPackage.dtsx"
D. dtexec /com "myPackage.dtsx"
Answer: C
Q74. You are completing the installation of the Data Quality Server component of SQL Server Data Quality Services (DQS).
You need to complete the post-installation configuration.
What should you do?
A. Install the Analysis Services OLE DB Provider.
B. Make the data available for DQS operations.
C. Run the Data Quality Server Installer.
D. Install ADOMD.NET.
Answer: C
Explanation:
References: http://msdn.microsoft.com/en-us/library/ff877917.aspx
http://msdn.microsoft.com/en-us/library/gg492277.aspx
Q75. To ease the debugging of packages, you standardize the SQL Server Integration Services (SSIS) package logging methodology.
The methodology has the following requirements:
Centralized logging in SQL Server
Simple deployment
Availability of log information through reports or T-SQL
Automatic purge of older log entries
Configurable log details
You need to configure a logging methodology that meets the requirements while minimizing the amount of deployment and development effort.
What should you do?
A. Deploy the package by using an msi file.
B. Use the gacutil command.
C. Create an OnError event handler.
D. Create a reusable custom logging component.
E. Use the dtutil /copy command.
F. Use the Project Deployment Wizard.
G. Run the package by using the dtexec /rep /conn command.
H. Add a data tap on the output of a component in the package data flow.
I. Run the package by using the dtexec /dumperror /conn command.
J. Run the package by using the dtexecui.exe utility and the SQL Log provider.
K. Deploy the package to the Integration Services catalog by using dtutil and use SQL Server to store the configuration.
Answer: I
Explanation: References: http://msdn.microsoft.com/en-us/library/ms140246.aspx
http://msdn.microsoft.com/en-us/library/hh231187.aspx
Q76. You are implementing a SQL Server Integration Services (SSIS) 2012 package that loads data from various flat files and a Windows Azure SQL Database database.
Daily transactions must be loaded into a staging database. All the SSIS tasks will use the CurrentDate variable as the transaction date.
You need to set the CurrentDate variable to the date stored in a control table of the Windows Azure SQL Database database when the package starts. You need to achieve this goal by using the least amount of development effort.
What should you use to set the variable?
A. an Expression task
B. an Execute SQL task
C. a Script component
D. a Script task
Answer: B
Q77. A SQL Server Integration Services (SSIS) project has been deployed to the SSIS catalog. The project includes a project Connection Manager to connect to the data warehouse.
The SSIS catalog includes two Environments:
. Development
. QA
Each Environment defines a single Environment Variable named ConnectionString of type string. The value of each variable consists of the connection string to the development or QA data warehouses.
You need to be able to execute deployed packages by using either of the defined Environments.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer:
Q78. You are using the Knowledge Discovery feature of the Data Quality Services (DQS) client application to modify an existing knowledge base.
In the mapping configuration, two of the three columns are mapped to existing domains in the knowledge base. The third column, named Group, does not yet have a domain.
You need to complete the mapping of the Group column.
What should you do?
A. Map a composite domain to the source column.
B. Create a composite domain that includes the Group column.
C. Add a domain for the Group column.
D. Add a column mapping for the Group column.
Answer: C
Q79. You manage a SQL Server Master Data Services (MDS) environment.
A new application requires access to the product data that is available in the MDS repository.
You need to design a solution that gives the application access to the product data with the least amount of development effort.
What should you do?
A. Use sp_addlinkedserver to add a linked server to access the MDS database tables directly.
B. Create an OLE DB connection string that sets the Provider property to MDS.
C. Use transactional replication for data synchronization.
D. Create a Subscription View in MDS.
Answer: D
Q80. A SQL Server Integration Services (SSIS) package named DataFeed interacts with an external vendor data feed. The package is executed several times a day, either as part of other packages' control flow or by itself. The external data feed is unreliable because network failures and slow response times are frequent. The package is currently deployed on the file system.
To analyze the reliability of the external data feed, you must collect execution data.
Every time the DataFeed package is executed, the following information must be logged:
Start Time
End Time
Execution Result
Execution Duration
You need to design a logging solution that meets the requirements by using the least amount of administrative and development effort.
Which three actions should you perform in sequence? (To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.)
Answer: