CS614-Midterm
1 / 50
De-Normalization normally speeds up
2 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
3 / 50
5 million bales.
4 / 50
NUMA stands for __________
5 / 50
Collapsing tables can be done on the ___________ relationships
6 / 50
When performing objective assessments, companies follow a set of principles to develop metrics specific to their needs, there is hard to have “one size fits all” approach. Which of the following statement represents the pervasive functional forms?
7 / 50
The growth of master files and magnetic tapes exploded around the mid- _______. :
8 / 50
A dense index, if fits into memory, costs only ______ disk I/O access to locate a record by given key.
9 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
10 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
11 / 50
Focusing on data warehouse delivery only often end up _________.
12 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
13 / 50
The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by _____________ tools typical of decision support systems.
14 / 50
For good decision making, data should be integrated across the organization to cross the LoB (Line of Business). This is to give the total view of organization from:
15 / 50
16 / 50
In a traditional MIS system, there is an almost linear sequence of queries.
17 / 50
Change Data Capture is one of the challenging technical issues in _____________
18 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
19 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
20 / 50
For a relation to be in 4NF it must be:-
21 / 50
The automated, prospective analyses offered by data mining move beyond the analysis of past events provided by respective tools typical of ___________.
22 / 50
If every key in the data file is represented in the index file then index is :
23 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
24 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
25 / 50
We must try to find the one access tool that will handle all the needs of their users.
26 / 50
For a DWH project, the key requirement are ________ and product experience.
27 / 50
To identify the __________________ required we need to perform data profiling
28 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
29 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
30 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
31 / 50
32 / 50
Investing years in architecture and forgetting the primary purpose of solving business problems, results in inefficient application. This is the example of _________ mistake.
33 / 50
Grain is the ________ level of data stored in the warehouse.
34 / 50
________ gives total view of an organization
35 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
36 / 50
Ad-hoc access means to run such queries which are known already.
37 / 50
The goal of star schema design is to simplify ________
38 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
39 / 50
Analytical processing uses ____________ , instead of record level access.
40 / 50
To judge effectiveness we perform data profiling twice.
41 / 50
DSS queries do not involve a primary key
42 / 50
43 / 50
The goal of ______is to look at as few block as possible to find the matching records.
44 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
45 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
46 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
47 / 50
Data mining derives its name from the similarities between searching for valuable business information in a large database, for example, finding linked products in gigabytes of store scanner data, and mining a mountain for a _________ of valuable ore.
48 / 50
If w is the window size and n is the size of data set, then the complexity of merging phase in BSN method is___________
49 / 50
Data mining is a/an ______ approach , where browsing through data using mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
50 / 50
Suppose the amount of data recorded in an organization is doubled every year. This increase is __________ .
Your score is
The average score is 0%
Restart quiz