CS614-Midterm
1 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
2 / 50
________ gives total view of an organization
3 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
4 / 50
_____ contributes to an under-utilization of valuable and expensive historical data, and inevitably results in a limitedcapability to provide decision support and analysis.
5 / 50
The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by _____________ tools typical of decision support systems.
6 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
7 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
8 / 50
.______ is class of Decision Support Environment.
9 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
10 / 50
_____________, if fits into memory , costs only one disk I/O access to locate a record by given key.
11 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
12 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
13 / 50
Rearranging the grouping of source data, delivering it to the destination database, and ensuring the quality of data are crucial to the process of loading the data warehouse. Data ____________ is vitally important to the overall health of a warehouse project. 1. Cleansing 2. Cleaning 3. Scrubbing Which of the following options is true?
14 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
15 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
16 / 50
Analytical processing uses ____________ , instead of record level access.
17 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
18 / 50
With data mining, the best way to accomplish this is by setting aside some of your data in a vault to isolate it from the mining process; once the mining is complete, the results can be tested against the isolated data to confirm the model's _______.
19 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
20 / 50
If some error occurs, execution will be terminated abnormally and all transactions will be rolled back. In this case when we will access the database we will find it in the state that was before the ____________.
21 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
22 / 50
_____modeling technique is more appropriate for data warehouses.
23 / 50
Focusing on data warehouse delivery only often end up _________.
24 / 50
NUMA stands for __________
25 / 50
The growth of master files and magnetic tapes exploded around the mid- _______. :
26 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
27 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
28 / 50
29 / 50
A dense index, if fits into memory, costs only ______ disk I/O access to locate a record by given key.
30 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
31 / 50
32 / 50
De-Normalization normally speeds up
33 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
34 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
35 / 50
In DWH project, it is assured that ___________ environment is similar to the production environment
36 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
37 / 50
Pre-computed _______ can solve performance problems
38 / 50
For a smooth DWH implementation we must be a technologist.
39 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
40 / 50
Ad-hoc access means to run such queries which are known already.
41 / 50
A single database, couldn‟t serve both operational high performance transaction processing and DSS, analytical processing, all at the same time.
42 / 50
43 / 50
Investing years in architecture and forgetting the primary purpose of solving business problems, results in inefficient application. This is the example of _________ mistake.
44 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
45 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
46 / 50
For a DWH project, the key requirement are ________ and product experience.
47 / 50
Data mining derives its name from the similarities between searching for valuable business information in a large database, for example, finding linked products in gigabytes of store scanner data, and mining a mountain for a _________ of valuable ore.
48 / 50
Cube is a __________ entity containing values of a certain fact at a certain aggregation level at an intersection of a combination of dimensions.
49 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The __________ the portion of the program that must be executed sequentially, the greater the scalability of computation.
50 / 50
When performing objective assessments, companies follow a set of principles to develop metrics specific to their needs, there is hard to have “one size fits all” approach. Which of the following statement represents the pervasive functional forms?
Your score is
The average score is 0%
Restart quiz