CS614-Midterm
1 / 50
Suppose the amount of data recorded in an organization is doubled every year. This increase is __________ .
2 / 50
Cube is a __________ entity containing values of a certain fact at a certain aggregation level at an intersection of a combination of dimensions.
3 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
4 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
5 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
6 / 50
Normalization effects performance
7 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
8 / 50
To judge effectiveness we perform data profiling twice.
9 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
10 / 50
NUMA stands for __________
11 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
12 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
13 / 50
All data is ______________ of something real. I An Abstraction II A Representation Which of the following option is true?
14 / 50
The need to synchronize data upon update is called
15 / 50
Slice and Dice is changing the view of the data.
16 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
17 / 50
5 million bales.
18 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
19 / 50
In a traditional MIS system, there is an almost linear sequence of queries.
20 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
21 / 50
Data warehousing and on-line analytical processing (OLAP) are _______ elements of decision support system.
22 / 50
The _________ is only a small part in realizing the true business value buried within the mountain of data collected and stored within organizations business systems and operational databases.
23 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
24 / 50
25 / 50
Grain is the ________ level of data stored in the warehouse.
26 / 50
Collapsing tables can be done on the ___________ relationships
27 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The __________ the portion of the program that must be executed sequentially, the greater the scalability of computation.
28 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
29 / 50
For a relation to be in 4NF it must be:-
30 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
31 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
32 / 50
When performing objective assessments, companies follow a set of principles to develop metrics specific to their needs, there is hard to have “one size fits all” approach. Which of the following statement represents the pervasive functional forms?
33 / 50
_______ is an application of information and data.
34 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
35 / 50
Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.
36 / 50
_________ breaks a table into multiple tables based upon common column values.
37 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
38 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
39 / 50
For good decision making, data should be integrated across the organization to cross the LoB (Line of Business). This is to give the total view of organization from:
40 / 50
The automated, prospective analyses offered by data mining move beyond the analysis of past events provided by respective tools typical of ___________.
41 / 50
Data mining is a/an ______ approach , where browsing through data using mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
42 / 50
The goal of star schema design is to simplify ________
43 / 50
Change Data Capture is one of the challenging technical issues in _____________
44 / 50
For a smooth DWH implementation we must be a technologist.
45 / 50
DSS queries do not involve a primary key
46 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
47 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
48 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
49 / 50
50 / 50
Analytical processing uses ____________ , instead of record level access.
Your score is
The average score is 0%
Restart quiz