CS614-Midterm
1 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
2 / 50
_____ contributes to an under-utilization of valuable and expensive historical data, and inevitably results in a limitedcapability to provide decision support and analysis.
3 / 50
_______ is an application of information and data.
4 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
5 / 50
DOLAP allows download of “cube” structures to a desktop platform with the need for shared relational or cube server.
6 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
7 / 50
If every key in the data file is represented in the index file then index is :
8 / 50
The need to synchronize data upon update is called
9 / 50
Focusing on data warehouse delivery only often end up _________.
10 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
11 / 50
The input to the data warehouse can come from OLTP or transactional system but not from other third party database.
12 / 50
A data warehouse implementation without an OLAP tool is always possible.
13 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The __________ the portion of the program that must be executed sequentially, the greater the scalability of computation.
14 / 50
5 million bales.
15 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
16 / 50
Investing years in architecture and forgetting the primary purpose of solving business problems, results in inefficient application. This is the example of _________ mistake.
17 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
18 / 50
_____________ is a process which involves gathering of information about column through execution of certain queries with intention to identify erroneous records.
19 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
20 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
21 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
22 / 50
_________ breaks a table into multiple tables based upon common column values.
23 / 50
Normalization effects performance
24 / 50
To judge effectiveness we perform data profiling twice.
25 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
26 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
27 / 50
28 / 50
A single database, couldn‟t serve both operational high performance transaction processing and DSS, analytical processing, all at the same time.
29 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
30 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
31 / 50
Pakistan is one of the five major ________ countries in the world.
32 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
33 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
34 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
35 / 50
If w is the window size and n is the size of data set, then the complexity of merging phase in BSN method is___________
36 / 50
37 / 50
38 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
39 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
40 / 50
De-Normalization normally speeds up
41 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
42 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
43 / 50
Pre-join technique is used to avoid
44 / 50
There are many variants of the traditional nested-loop join. If the index is built as part of the query plan and subsequently dropped, it is called
45 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
46 / 50
B-Tree is used as an index to provide access to records
47 / 50
Slice and Dice is changing the view of the data.
48 / 50
Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.
49 / 50
With data mining, the best way to accomplish this is by setting aside some of your data in a vault to isolate it from the mining process; once the mining is complete, the results can be tested against the isolated data to confirm the model's _______.
50 / 50
Companies collect and record their own operational data, but at the same time they also use reference data obtained from _______ sources such as codes, prices etc.
Your score is
The average score is 0%
Restart quiz