CS614-Midterm
1 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
2 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
3 / 50
Data mining evolve as a mechanism to cater the limitations of ________ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc.
4 / 50
We must try to find the one access tool that will handle all the needs of their users.
5 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
6 / 50
Node of a B-Tree is stored in memory block and traversing a B-Tree involves ______ page faults.
7 / 50
If w is the window size and n is the size of data set, then the complexity of merging phase in BSN method is___________
8 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
9 / 50
: The goal of ___________ is to look at as few blocks as possible to find the matching records(s).
10 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
11 / 50
Normalization effects performance
12 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
13 / 50
_____________ is a process which involves gathering of information about column through execution of certain queries with intention to identify erroneous records.
14 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
15 / 50
The goal of ______is to look at as few block as possible to find the matching records.
16 / 50
Change Data Capture is one of the challenging technical issues in _____________
17 / 50
To judge effectiveness we perform data profiling twice.
18 / 50
19 / 50
20 / 50
Pakistan is one of the five major ________ countries in the world.
21 / 50
Pre-computed _______ can solve performance problems
22 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
23 / 50
For a DWH project, the key requirement are ________ and product experience.
24 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
25 / 50
26 / 50
In DWH project, it is assured that ___________ environment is similar to the production environment
27 / 50
Ad-hoc access means to run such queries which are known already.
28 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
29 / 50
Which statement is true for De-Normalization?
30 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
31 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
32 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
33 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
34 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
35 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
36 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
37 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
38 / 50
For a given data set, to get a global view in un-supervised learning we use
39 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
40 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
41 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
42 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
43 / 50
De-Normalization normally speeds up
44 / 50
As apposed to the out come of classification, estimation deal with ____________ valued outcome.
45 / 50
Data mining derives its name from the similarities between searching for valuable business information in a large database, for example, finding linked products in gigabytes of store scanner data, and mining a mountain for a _________ of valuable ore.
46 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
47 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
48 / 50
_______ is an application of information and data.
49 / 50
Transactional fact tables do not have records for events that do not occur. These are called
50 / 50
Your score is
The average score is 0%
Restart quiz