CS614-Midterm
1 / 50
For a DWH project, the key requirement are ________ and product experience.
2 / 50
_______________, if fits into memory, costs only one disk I/O access to locate a record by given key.
3 / 50
NUMA stands for __________
4 / 50
The Kimball s iterative data warehouse development approach drew on decades of experience to develop the _________.
5 / 50
Change Data Capture is one of the challenging technical issues in _____________
6 / 50
Multidimensional databases typically use proprietary __________ format to store pre-summarized cube structures.
7 / 50
Ad-hoc access means to run such queries which are known already.
8 / 50
DOLAP allows download of “cube” structures to a desktop platform with the need for shared relational or cube server.
9 / 50
Data Warehouse provides the best support for analysis while OLAP carries out the _________ task.
10 / 50
Rearranging the grouping of source data, delivering it to the destination database, and ensuring the quality of data are crucial to the process of loading the data warehouse. Data ____________ is vitally important to the overall health of a warehouse project. 1. Cleansing 2. Cleaning 3. Scrubbing Which of the following options is true?
11 / 50
Focusing on data warehouse delivery only often end up _________.
12 / 50
If some error occurs, execution will be terminated abnormally and all transactions will be rolled back. In this case when we will access the database we will find it in the state that was before the ____________.
13 / 50
The performance in a MOLAP cube comes from the O(1) look-up time for the array data structure.
14 / 50
5 million bales.
15 / 50
If w is the window size and n is the size of data set, then the complexity of merging phase in BSN method is___________
16 / 50
Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.
17 / 50
________ gives total view of an organization
18 / 50
Data mining is a/an __________ approach, where browsing through data using data mining techniques may reveal something that might be of interest to the user as information that was unknown previously.
19 / 50
We must try to find the one access tool that will handle all the needs of their users.
20 / 50
Grain is the ________ level of data stored in the warehouse.
21 / 50
Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model.
22 / 50
_________ breaks a table into multiple tables based upon common column values.
23 / 50
The input to the data warehouse can come from OLTP or transactional system but not from other third party database.
24 / 50
A data warehouse implementation without an OLAP tool is always possible.
25 / 50
26 / 50
Many data warehouse project teams waste enormous amounts of time searching in vain for a _______.
27 / 50
Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as
28 / 50
The goal of star schema design is to simplify ________
29 / 50
The key idea behind ___________ is to take a big task and break it into subtasks that can be processed concurrently on a stream of data inputs in multiple, overlapping stages of execution.
30 / 50
31 / 50
The purpose of the House of Quality technique is to reduce ______ types of risk.
32 / 50
In _________ system, the contents change with time. :
33 / 50
Normalization effects performance
34 / 50
A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.
35 / 50
Data mining uses _________ algorithms to discover patterns and regularities in data.
36 / 50
Data Transformation Services (DTS) provide a set of _____ that lets you extract, transform, and consolidate data from disparate sources into single or multiple destinations supported by DTS connectivity.
37 / 50
Data mining evolve as mechanism to cater the limitations of _____ systems to deal massive data sets with high dimensionality, new data types, multiple heterogeneous data resources etc...
38 / 50
39 / 50
To identify the __________________ required we need to perform data profiling
40 / 50
____________ in agriculture extension is that pest population beyond which the benefit of spraying outweighs its cost.
41 / 50
Pre-computed _______ can solve performance problems
42 / 50
The divide & conquer cube partitioning approach helps alleviate the ____________ limitations of MOLAP implementation.
43 / 50
_____ contributes to an under-utilization of valuable and expensive historical data, and inevitably results in a limitedcapability to provide decision support and analysis.
44 / 50
Pipeline parallelism focuses on increasing throughput of task execution, NOT on __________ sub-task execution time.
45 / 50
The _________ is only a small part in realizing the true business value buried within the mountain of data collected and stored within organizations business systems and operational databases.
46 / 50
The technique that is used to perform these feats in data mining modeling, and this act of model building is something that people have been doing for long time, certainly before the _______ of computers or data mining technology.
47 / 50
Analytical processing uses ____________ , instead of record level access.
48 / 50
De-Normalization normally speeds up
49 / 50
The degree of similarity between two records, often measured by a numerical value between _______, usually depends on application characteristics.
50 / 50
A single database, couldn‟t serve both operational high performance transaction processing and DSS, analytical processing, all at the same time.
Your score is
The average score is 0%
Restart quiz