CS614-Midterm
1 / 50
Pre-join technique is used to avoid
2 / 50
Suppose the amount of data recorded in an organization is doubled every year. This increase is __________ .
3 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
4 / 50
________ gives total view of an organization
5 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
6 / 50
DSS queries do not involve a primary key
7 / 50
The need to synchronize data upon update is called
8 / 50
Pre-computed _______ can solve performance problems
9 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
10 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
11 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
12 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
13 / 50
Change Data Capture is one of the challenging technical issues in _____________
14 / 50
Analytical processing uses ____________ , instead of record level access.
15 / 50
If every key in the data file is represented in the index file then index is :
16 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
17 / 50
The goal of star schema design is to simplify ________
18 / 50
As apposed to the out come of classification, estimation deal with ____________ valued outcome.
19 / 50
Transactional fact tables do not have records for events that do not occur. These are called
20 / 50
_____ contributes to an under-utilization of valuable and expensive historical data, and inevitably results in a limitedcapability to provide decision support and analysis.
21 / 50
For a DWH project, the key requirement are ________ and product experience.
22 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
23 / 50
B-Tree is used as an index to provide access to records
24 / 50
For good decision making, data should be integrated across the organization to cross the LoB (Line of Business). This is to give the total view of organization from:
25 / 50
With data mining, the best way to accomplish this is by setting aside some of your data in a vault to isolate it from the mining process; once the mining is complete, the results can be tested against the isolated data to confirm the model's _______.
26 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
27 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
28 / 50
29 / 50
For a given data set, to get a global view in un-supervised learning we use
30 / 50
NUMA stands for __________
31 / 50
When performing objective assessments, companies follow a set of principles to develop metrics specific to their needs, there is hard to have “one size fits all” approach. Which of the following statement represents the pervasive functional forms?
32 / 50
_____________, if fits into memory , costs only one disk I/O access to locate a record by given key.
33 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
34 / 50
In DWH project, it is assured that ___________ environment is similar to the production environment
35 / 50
Slice and Dice is changing the view of the data.
36 / 50
37 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
38 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
39 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
40 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
41 / 50
: The goal of ___________ is to look at as few blocks as possible to find the matching records(s).
42 / 50
.______ is class of Decision Support Environment.
43 / 50
Collapsing tables can be done on the ___________ relationships
44 / 50
The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by _____________ tools typical of decision support systems.
45 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
46 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
47 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
48 / 50
49 / 50
in agriculture extension is that pest population beyond which the benefit of spraying outweighs levels
50 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
Your score is
The average score is 0%
Restart quiz