CS614-Midterm
1 / 50
In _________ system, the contents change with time. :
2 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
3 / 50
It is observed that every year the amount of data recorded in an organization :
4 / 50
Focusing on data warehouse delivery only often end up _________.
5 / 50
6 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
7 / 50
Which statement is true for De-Normalization?
8 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
9 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
10 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
11 / 50
Ad-hoc access means to run such queries which are known already.
12 / 50
Analytical processing uses ____________ , instead of record level access.
13 / 50
14 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
15 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
16 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
17 / 50
Change Data Capture is one of the challenging technical issues in _____________
18 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
19 / 50
Data Warehouse provides the best support for analysis while OLAP carries out the _________ task.
20 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
21 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
22 / 50
The need to synchronize data upon update is called
23 / 50
_____________, if fits into memory , costs only one disk I/O access to locate a record by given key.
24 / 50
Rearranging the grouping of source data, delivering it to the destination database, and ensuring the quality of data are crucial to the process of loading the data warehouse. Data ____________ is vitally important to the overall health of a warehouse project. 1. Cleansing 2. Cleaning 3. Scrubbing Which of the following options is true?
25 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
26 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
27 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
28 / 50
A data warehouse implementation without an OLAP tool is always possible.
29 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
30 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
31 / 50
If every key in the data file is represented in the index file then index is :
32 / 50
To identify the __________________ required we need to perform data profiling
33 / 50
The _________ is only a small part in realizing the true business value buried within the mountain of data collected and stored within organizations business systems and operational databases.
34 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
35 / 50
The users of data warehouse are knowledge workers in other words they are _______in the organization.
36 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
37 / 50
Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.
38 / 50
____________ in agriculture extension is that pest population beyond which the benefit of spraying outweighs its cost.
39 / 50
Data mining is a/an ______ approach , where browsing through data using mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
40 / 50
Pre-join technique is used to avoid
41 / 50
.______ is class of Decision Support Environment.
42 / 50
We must try to find the one access tool that will handle all the needs of their users.
43 / 50
If „M‟ rows from table-A match the conditions in the query then table-B is accessed „M‟ times. Suppose table-B has an index on the join column. If „a‟ I/Os are required to read the data block for each scan and „b‟ I/Os for each data block then the total cost of accessing table-B is _____________ logical I/Os approximately.
44 / 50
45 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
46 / 50
Pre-computed _______ can solve performance problems
47 / 50
The input to the data warehouse can come from OLTP or transactional system but not from other third party database.
48 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
49 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
50 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
Your score is
The average score is 0%
Restart quiz