CS614-Midterm
1 / 50
During ETL process of an organization, suppose you have data which can be transformed using any of the transformation method. Which of the following strategy will be your choice for least complexity?
2 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
3 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
4 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
5 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
6 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
7 / 50
A single database, couldn‟t serve both operational high performance transaction processing and DSS, analytical processing, all at the same time.
8 / 50
Horizontal splitting breaks a table into multiple tables based upon_______
9 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
10 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
11 / 50
Data mining derives its name from the similarities between searching for valuable business information in a large database, for example, finding linked products in gigabytes of store scanner data, and mining a mountain for a _________ of valuable ore.
12 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
13 / 50
The users of data warehouse are knowledge workers in other words they are _______in the organization.
14 / 50
For a DWH project, the key requirement are ________ and product experience.
15 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
16 / 50
Grain is the ________ level of data stored in the warehouse.
17 / 50
A data warehouse implementation without an OLAP tool is always possible.
18 / 50
in agriculture extension is that pest population beyond which the benefit of spraying outweighs levels
19 / 50
Pre-computed _______ can solve performance problems
20 / 50
Which statement is true for De-Normalization?
21 / 50
Cube is a __________ entity containing values of a certain fact at a certain aggregation level at an intersection of a combination of dimensions.
22 / 50
For a relation to be in 4NF it must be:-
23 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
24 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
25 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
26 / 50
DOLAP allows download of “cube” structures to a desktop platform with the need for shared relational or cube server.
27 / 50
28 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
29 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
30 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
31 / 50
Focusing on data warehouse delivery only often end up _________.
32 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
33 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
34 / 50
_____modeling technique is more appropriate for data warehouses.
35 / 50
36 / 50
Change Data Capture is one of the challenging technical issues in _____________
37 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
38 / 50
De-Normalization normally speeds up
39 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The __________ the portion of the program that must be executed sequentially, the greater the scalability of computation.
40 / 50
Collapsing tables can be done on the ___________ relationships
41 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
42 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
43 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
44 / 50
We must try to find the one access tool that will handle all the needs of their users.
45 / 50
_________ breaks a table into multiple tables based upon common column values.
46 / 50
The input to the data warehouse can come from OLTP or transactional system but not from other third party database.
47 / 50
48 / 50
Normalization effects performance
49 / 50
Data mining is a/an ______ approach , where browsing through data using mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
50 / 50
Your score is
The average score is 0%
Restart quiz