CS614-Midterm
1 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
2 / 50
Collapsing tables can be done on the ___________ relationships
3 / 50
It is observed that every year the amount of data recorded in an organization :
4 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
5 / 50
The input to the data warehouse can come from OLTP or transactional system but not from other third party database.
6 / 50
____________ in agriculture extension is that pest population beyond which the benefit of spraying outweighs its cost.
7 / 50
Slice and Dice is changing the view of the data.
8 / 50
There are many variants of the traditional nested-loop join, if there is an index is exploited, then it is called……
9 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
10 / 50
Cube is a __________ entity containing values of a certain fact at a certain aggregation level at an intersection of a combination of dimensions.
11 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
12 / 50
Pakistan is one of the five major ________ countries in the world.
13 / 50
If some error occurs, execution will be terminated abnormally and all transactions will be rolled back. In this case when we will access the database we will find it in the state that was before the ____________.
14 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
15 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
16 / 50
In DWH project, it is assured that ___________ environment is similar to the production environment
17 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
18 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
19 / 50
_______________, if fits into memory, costs only one disk I/O access to locate a record by given key.
20 / 50
Transactional fact tables do not have records for events that do not occur. These are called
21 / 50
Change Data Capture is one of the challenging technical issues in _____________
22 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
23 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
24 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
25 / 50
NUMA stands for __________
26 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
27 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
28 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The __________ the portion of the program that must be executed sequentially, the greater the scalability of computation.
29 / 50
For a given data set, to get a global view in un-supervised learning we use
30 / 50
As apposed to the out come of classification, estimation deal with ____________ valued outcome.
31 / 50
During ETL process of an organization, suppose you have data which can be transformed using any of the transformation method. Which of the following strategy will be your choice for least complexity?
32 / 50
Data Warehouse provides the best support for analysis while OLAP carries out the _________ task.
33 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
34 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
35 / 50
If „M‟ rows from table-A match the conditions in the query then table-B is accessed „M‟ times. Suppose table-B has an index on the join column. If „a‟ I/Os are required to read the data block for each scan and „b‟ I/Os for each data block then the total cost of accessing table-B is _____________ logical I/Os approximately.
36 / 50
If w is the window size and n is the size of data set, then the complexity of merging phase in BSN method is___________
37 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
38 / 50
The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by _____________ tools typical of decision support systems.
39 / 50
Data warehousing and on-line analytical processing (OLAP) are _______ elements of decision support system.
40 / 50
The growth of master files and magnetic tapes exploded around the mid- _______. :
41 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
42 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
43 / 50
The goal of star schema design is to simplify ________
44 / 50
To identify the __________________ required we need to perform data profiling
45 / 50
_____modeling technique is more appropriate for data warehouses.
46 / 50
Analytical processing uses ____________ , instead of record level access.
47 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
48 / 50
________ gives total view of an organization
49 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
50 / 50
Normalization effects performance
Your score is
The average score is 0%
Restart quiz