CS614-Midterm
1 / 50
DTS allows us to connect through any data source or destination that is supported by ____________
2 / 50
To judge effectiveness we perform data profiling twice.
3 / 50
For a given data set, to get a global view in un-supervised learning we use
4 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
5 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
6 / 50
As apposed to the out come of classification, estimation deal with ____________ valued outcome.
7 / 50
Collapsing tables can be done on the ___________ relationships
8 / 50
If „M‟ rows from table-A match the conditions in the query then table-B is accessed „M‟ times. Suppose table-B has an index on the join column. If „a‟ I/Os are required to read the data block for each scan and „b‟ I/Os for each data block then the total cost of accessing table-B is _____________ logical I/Os approximately.
9 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
10 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
11 / 50
_________ breaks a table into multiple tables based upon common column values.
12 / 50
________ gives total view of an organization
13 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
14 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
15 / 50
In _________ system, the contents change with time. :
16 / 50
Horizontal splitting breaks a table into multiple tables based upon_______
17 / 50
DSS queries do not involve a primary key
18 / 50
Companies collect and record their own operational data, but at the same time they also use reference data obtained from _______ sources such as codes, prices etc.
19 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
20 / 50
The goal of ideal parallel execution is to completely parallelize those parts of a computation that are not constrained by data dependencies. The smaller the portion of the program that must be executed __________, the greater the scalability of the computation.
21 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
22 / 50
The need to synchronize data upon update is called
23 / 50
24 / 50
: The goal of ___________ is to look at as few blocks as possible to find the matching records(s).
25 / 50
Grain is the ________ level of data stored in the warehouse.
26 / 50
_____modeling technique is more appropriate for data warehouses.
27 / 50
De-Normalization normally speeds up
28 / 50
During the application specification activity, we also must give consideration to the organization of the applications.
29 / 50
Analytical processing uses ____________ , instead of record level access.
30 / 50
If some error occurs, execution will be terminated abnormally and all transactions will be rolled back. In this case when we will access the database we will find it in the state that was before the ____________.
31 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
32 / 50
Suppose the amount of data recorded in an organization is doubled every year. This increase is __________ .
33 / 50
Normalization effects performance
34 / 50
All data is ______________ of something real. I An Abstraction II A Representation Which of the following option is true?
35 / 50
36 / 50
Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.
37 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
38 / 50
B-Tree is used as an index to provide access to records
39 / 50
_______________, if too big and does not fit into memory, will be expensive when used to find a record by given key.
40 / 50
The goal of ______is to look at as few block as possible to find the matching records.
41 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
42 / 50
Change Data Capture is one of the challenging technical issues in _____________
43 / 50
In DWH project, it is assured that ___________ environment is similar to the production environment
44 / 50
The goal of star schema design is to simplify ________
45 / 50
Which statement is true for De-Normalization?
46 / 50
During ETL process of an organization, suppose you have data which can be transformed using any of the transformation method. Which of the following strategy will be your choice for least complexity?
47 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
48 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
49 / 50
Pre-computed _______ can solve performance problems
50 / 50
For a relation to be in 4NF it must be:-
Your score is
The average score is 0%
Restart quiz