CS614-Midterm
1 / 50
The need to synchronize data upon update is called
2 / 50
Horizontal splitting breaks a table into multiple tables based upon_______
3 / 50
NUMA stands for __________
4 / 50
De-Normalization normally speeds up
5 / 50
Multi-dimensional databases (MDDs) typically use ___________ formats to store pre-summarized cube structures.
6 / 50
To measure or quantify the similarity or dissimilarity, different techniques are available. Which of the following option represent the name of available techniques?
7 / 50
Normalization effects performance
8 / 50
Non uniform distribution, when the data is distributed across the processors, is called ______.
9 / 50
DOLAP allows download of “cube” structures to a desktop platform with the need for shared relational or cube server.
10 / 50
In horizontal splitting, we split a relation into multiple tables on the basis of
11 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
12 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
13 / 50
_____modeling technique is more appropriate for data warehouses.
14 / 50
If „M‟ rows from table-A match the conditions in the query then table-B is accessed „M‟ times. Suppose table-B has an index on the join column. If „a‟ I/Os are required to read the data block for each scan and „b‟ I/Os for each data block then the total cost of accessing table-B is _____________ logical I/Os approximately.
15 / 50
If some error occurs, execution will be terminated abnormally and all transactions will be rolled back. In this case when we will access the database we will find it in the state that was before the ____________.
16 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
17 / 50
Execution can be completed successfully or it may be stopped due to some error. In case of successful completion of execution all the transactions will be ___________
18 / 50
If every key in the data file is represented in the index file then index is :
19 / 50
20 / 50
Rearranging the grouping of source data, delivering it to the destination database, and ensuring the quality of data are crucial to the process of loading the data warehouse. Data ____________ is vitally important to the overall health of a warehouse project. 1. Cleansing 2. Cleaning 3. Scrubbing Which of the following options is true?
21 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
22 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
23 / 50
The automated, prospective analyses offered by data mining move beyond the analyses of past events provided by _____________ tools typical of decision support systems.
24 / 50
_______ is an application of information and data.
25 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
26 / 50
Analytical processing uses ____________ , instead of record level access.
27 / 50
A single database, couldn‟t serve both operational high performance transaction processing and DSS, analytical processing, all at the same time.
28 / 50
With data mining, the best way to accomplish this is by setting aside some of your data in a vault to isolate it from the mining process; once the mining is complete, the results can be tested against the isolated data to confirm the model's _______.
29 / 50
Collapsing tables can be done on the ___________ relationships
30 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
31 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
32 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
33 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
34 / 50
35 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
36 / 50
_________ breaks a table into multiple tables based upon common column values.
37 / 50
: An optimized structure which is built primarily for retrieval, with update being only a secondary consideration is
38 / 50
Slice and Dice is changing the view of the data.
39 / 50
For good decision making, data should be integrated across the organization to cross the LoB (Line of Business). This is to give the total view of organization from:
40 / 50
It is observed that every year the amount of data recorded in an organization :
41 / 50
________ is the technique in which existing heterogeneous segments are reshuffled, relocated into homogeneous segments.
42 / 50
_____________, if fits into memory , costs only one disk I/O access to locate a record by given key.
43 / 50
44 / 50
The STAR schema used for data design is a __________ consisting of fact and dimension tables. :
45 / 50
If someone told you that he had a good model to predict customer usage, the first thing you might try would be to ask him to apply his model to your customer _______, where you already knew the answer.
46 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
47 / 50
30.Data Warehouse is about taking / colleting data from different ________ sources:
48 / 50
People that design and build the data warehouse must be capable of working across the organization at all levels
49 / 50
We must try to find the one access tool that will handle all the needs of their users.
50 / 50
Virtual cube is used to query two similar cubes by creating a third “virtual” cube by a join between two cubes.
Your score is
The average score is 0%
Restart quiz