VU Grand Quiz

CS614 Solved Grand Quiz Spring 2021

CS614 Solved Grand Quiz

Welcome to CS614 Grand Quizzes at VUExamGuide.com – your ultimate destination for 100% Correct Answers! If you are a CS614 student at VU, looking for a reliable study resource to conquer your quizzes and exams, you’ve come to the right place.

Our CS614 Grand Quizzes cover a wide range of topics, meticulously prepared by our expert team to ensure accuracy and depth of understanding. With a focus on essential computer science concepts, these quizzes will sharpen your problem-solving skills and strengthen your grasp of advanced algorithms and data structures.

Say goodbye to exam stress and uncertain solutions! VUExamGuide.com empowers you to excel in CS614. Embrace the knowledge and skills you need to tackle complex challenges and achieve remarkable success in your course.

Join us now and embark on a journey of academic excellence with CS614 Grand Quizzes. Prepare to shine and propel yourself towards a brighter future – start exploring VUExamGuide.com today!

CS614 SOLVED MCQs

Onlline Extraction is a kind of———————————data extraction.

  • Logical
  • Dimensional
  • Physical page 132
  • Multi valued

The —————- saw the advent of disk storage, or DASD( direct Access Storage Device) :

  • 1960s

•    1970s page 13

  • 1950s
  • 1990s

In context of data warehouse, normally it becomes difficult to extract data from different sources because these sources are normally.

•    Heterogeneous       page 140

  • Homogeneous
  • Centralized
  • Baseline

Which of the following is not a task of Data Transformation?

  • Conversion
  • Summarization
  • Enrichment

•    Full Data Refresh  page 135

Which of the following is not an Orr’s Law of Data Quality”?

  • “Data that is not used cannot be corrected!”
  • “Data quality is a function of its use, not its collection!”
  • “Data will be no better than its most stringent use!”

•    “Data duplication can be harmful for the organization! ”      page 181

Flat files are one of the prevalent structures used in ——————- data extraction:

  • Online

•    Offline            page 134

  • Incremental
  • Full

Which of the following is NOT one of the advantages of changed data capture (CDC) technique?

  • Flat files are not required

•    Limited query interface is required for data extraction           page 152

  • No incremental on-line I/O required for log tape
  • Extraction of changed data occurs immediately

The most common range partitioning is on

  • Color

•    Date         page 66

  • PhoneNo
  • Name

A relation is said to be in first normal form(1NF), if it does not contain ________

  • Single value column

•    Multi-valued column      page 43

  • Derived column
  • Composite column

In a fully normalized database, too many ____________are required

  • Values

•    Joins                 page 49

  • Queries
  • Conditions

In the data warehouse, data is collection from ——————– sources:

  • Homogeneous

•    Heterogeneous                  page 21

  • External
  • Internal

De-normalization is more like a “controlled crash” with the aim to ———— without loss of information:

  • Check
  • Balance
  • Decrease

•    Enhance    page 49

—————– is making all efforts to increase effectiveness and efficiency in meeting and accepted customer expectation:

  • Quality assurance

   Quality improvement   page 183

  • Quality maintenance
  • Quality Establishment

————- is the application of intelligence and experience to get common goals.

   Wisdom        page 11

  • Education
  • Power
  • Information

In the data transformation, ———- is the rearrangement and simplification of individual

  • Aggregation
  • Enrichment  page 136
  • Splitting joining
  • Conversion

Grain of a fact table means :

   The meaning of one fact table row   page 109

  • The meaning of one dimensional table row
  • Summary of aggregates in all fact tables
  • Summary of aggregates in all dimension tables

Normalization —————– :

   Reduces redundancy          page 41

  • Increases redundancy
  • Reduces joins
  • Reduces tables

Which of the following is not an example of a typical grain :

  • Individual transaction
  • Daily aggregates
  • Monthly aggregates

•    Normalized attributes    page 111

Multi-dimensional databases(MDDs) typically use ——————– formats to store pre-summarized cube structures:

  • SQL

•    Proprietary file          page 79

  • Object oriented
  • Non-proprietary file

———— provides a combination of “relational databases access” and “cube” data structures within a single framework:

•    HOLAP                       page 78

  • DOLAP
  • MOLAP
  • ROLAP

Data Warehouse provides the best support for analysis while OLAP carries out the ————-

task:

  • Mandatory
  • Whole

•    Analysis                                page 69

  • Prediction

—————— involves splitting a table by columns so that a group of columns is placed into the new table and  the remaining columns are placed in another new table:

•    Vertical splitting             page 56

  • Horizontal splitting
  • Adding redundant column
  • None of the given option

OLAP implementations are highly/completely —————— :

  • Normalized

•    Demoralized             page 69

  • Predictive
  • Additive

If each cell of Relation R contains a single value ( no repeating values) then it is confirmed that :

•    Relation R is in 1st Normal Form       page 43

  • Relation R is in 2nd Normal Form
    • Relation R is in 3rd Normal Form
    • Relation R is in 3rd Normal Form but not in 2nd Normal Form

Which kind of relationships is captured by Fact less fact table:

•    Many- to- Many                  page 121

  • One-to-many
  • One-to-one
  • None of the given option

Which of the following is not an example of dimension:

  • Product
  • Date
  • Region

•    Sales volume                        page 78

Which people criticize Dimensional Modeling (DM) as being a data mart oriented approach?

  • Those that consider ER models as Data marts

•    Those that consider Business processes as Data marts  page 110

  • Those that consider Data marts as Data warehouse
  • Those that consider dimensional model
  • Those that consider dimensional modeling as de-normalization approach

In a fully normalized form:

•    To many joins are required         page 49

  • Relationships lose their significance
  • No joins are required
  • Data integrity becomes an issue

Which of the following is an example of Non-Additive Facts:

  • Quality sold
  • Total sale in Rs.

•    Discount in percentage                        page 119

  • Count of orders in a store

Which of the following is not a CUBE operation?

•    ANSI SQL       page 81

  • Roll Up
  • Drill Down
  • Pivoting

——————– allows download of “cube” structures to a desktop platform without the need for shared or cube server:

  • MPLAP
  • ROLAP

•    DOLAP                                    page 78

  • HOLAP

ROLAP provides access to information via a relational database using:

•    ANSI standard SQL                                        page 78

  • Proprietary file format
  • Comma Separated Values
  • All of the given option

——————– is usually deployed when expression can be used to group data together in such a way that access can be targeted to a small set of partitions:

  • Expression elimination

•    Expression partitioning                                 page 67

  • •         Expression indexing
  • None of the given option

Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as

  • Distributed Systems Architecture

•    Legacy System Architecture                                 page 14

  • Online System Architecture
  • Intranet System Architecture

The data has to be checked , cleaned and transformed into a ————— format to allow easy and fast access

•    Unified                                                                 page 20

  • Predicated
  • Qualified
  • Proactive

Suppose in a system A, the values of “PhoneNo” attribute were stored in “countrycode-phone-extension” format, however after transformation into data warehouse the separate columns were used for “countrycode”,”phone” and “extension”. The above scenario is an example of :

  • One-to-one scalar transformation

•    One-to-many element transformation         page 144+conceptual

  • Many-to-one element transformation
  • Many-to-many element transformation

In decision support system ease of use in achieved by:

  • Normalization

•    Denormalization                     page no 49

  • Drill up
  • Drill down

Which of the following is one of the methods to simplify an ER model?

  • Normalization

•    Denormalization                           page no 103

  • HOLAP
  • Hybrid schema

 In ETL process data transformation includes —————-

•    Data cleansing                                        page 129

  • Data aggregation
  • Behavior checking
  • Pattern recognition

Non-uniform use of abbreviations, units, and values refers to:

•    Syntactically dirty data                       page 160

  • Semantically dirty data
  • Coverage anomaly
  • Extraction issue

Suppose the size of the attribute “Computerized National Card (CNIC) no. is changed in NADRA database. This transformation refers to:

•    Format revision                                           page 153

  • Field splitting
  • Field decoding
  • Calculation of derived value

The divide and conquer cube partitioning approach helps alleviate the ———— limitations of MOLAP implementation:

  • Flexibility
  • Maintainability
  • Security

•    Scalability                                                 page 85

identify the TRUE statement:

  • DM is inherently dimensional in nature
  • DM comprises of a single central fact table
  • DM comprises of a set of dimensional tables

•    All of the given option                               Page 103

————- can be used when some columns are rarely accessed rather than other columns or when the table has wide rows or header or both:

  • Horizontal splitting
  • Pre-joining

•    Vertical splitting                               page 56

  • Derived attributes

Which of the following is an example of derived attributes?

•    Age                                                                    page 61

  • Size
  • Color
  • Length

The online high performance transaction processing was evolved in ————–:

  • 1980

•    1975                                                   page 12

  • 1977
  • 1965

Cube is a logical entity containing values of a certain aggregation level at an intersection of a combination of ——————– :

  • Facts

•    Dimension                                            page 88

  • Summary tables
  • Primary and foreign key

Which of the following is TRUE regarding Entity relationship modeling?

  • It does not really model business, but models the micro relationships among data elements.
  • ER modeling does not have “business rules,” it has “data rules
  • ER modeling helps retrieval of individual records having certain critical identifiers.

•    All of the given option                          page 102

——Facilitates a mobile computing paradiagramn:

  • HOLAP
    • DOLAP               page78
  • ROLAP
  • MOLAP

The main reason(s )for the increase in cube size may be:

  • Increase in the number of dimensions
  • Increase in the cardinality of the dimensions
  • Increase in the amount of detail data

•    All of the given options page 87

Suppose the amount of data recorded in an organization is doubled in year. This increase in ——

  • Linear
  • Quadratic

•    Exponential                            page 15

  • Logarithmic

The data in the data warehouse is ———– :

  • Volatile

•    Non-volatile                         page 69

  • Static
  • Non-structured

————— models the macro relationships among data elements with an overall deterministic strategy:

•    Dimensional model              page102

  • Entity relationship model
  • Object oriented model
  • Structured model

—————– technique requires a separate column to specify the time and date when the last modification was occurred:

  • Checkmarks

•    Timestamps           page 150

  • Just-in-Time
  • Real Time extraction

Which of the de-normalization technique squeezes master table into detail?

•    Pre-joining                     page 58

  • Horizontal splitting
  • Vertical splitting
  • Adding redundant column

De-normalization can help:

  • Minimize joins
  • Minimize foreign keys
  • Resolve aggregates

•    All of the given options page 51

The domain of the “gender” field in some database may be (‘F’,’M’) or as (“Female”, “Male”) or even as (1, 0). This is:

  • Primary key problem

•    Non primary key problem        page 163

  • Normalization problem
  • All of the given option
 Increasing level of normalization —————-number of tables:
    
 •  Increasespage 51 
  • Decreases
  • Does not effect
  • None of the given option

Which of the following is not a Data Quality Validation Technique:

  • Referential integrity
  • Using Data Quality Rules
  • Data Histograming

•    Indexes                    page 189

This technique can be used when column from one table is frequently accessed in a large scale join in conjunction with a column from another table:

  • Horizontal splitting
  • Pre-joining

•    Adding redundant column page 58

  • Derived attributes

Data cleansing requires involvement of domain expert because:

  • Domain expert has deep knowledge of data aggregation
  • Change Data captures requires involvement of domain expert

•    Domain knowledge is required to correct anomalies     page 158

  • Domain expert has deep knowledge of data summarization

Relational databases allow you to navigate the data in ————- that is appropriate using the primary , foreign key structure with in the data model:

  • Only One Direction

•    Any Direction       page 19

  • Two Direction
  • None of these

History is excellent predicator of the ————:

  • Past
  • Present

•    Future                   page 15

  • History

De- normalization is the process of selectively transforming normalized relations into un-normalized physical record specifications, with the aim to:

  • Well structure the data
  • Well model the data

•    Reduce query processing time page 50

  • None of the given option

—————– gives total view of an organization:

  • OLAP
  • OLTP

   Data Warehouse     page 16

  • Database

Suppose in system A, the possible values of “Gender” attribute were “Male”& “Female”, however in data warehouse ,the values stored were “M” for male and “F” for female. This above scenario is an example of :

   One-to-one scalar transformation      page 144

  • One-to-many element transformation
  • Many-to-one element transformation
  • Many-to-many element transformation

Enrichment is one of the basic tasks in data —————- :

  • Extraction

•    Transformation page 138

  • Loading
  • Summarization

Which of the following is not a technique of De-normalization?

  • Pre-joining
  • Splitting tables
  • Adding redundant columns

•    ER modeling                  page 52

Which of the following is an example of Additive Facts?

•    Sales Amount              page 119

  • Average
  • Discount
  • Ratios

Robotic libraries are needed for ————————-:

  • Cubes
  • Data marts

•    Data warehouse                page 131

  • Aggregates

Normally ROLAP is implemented using —————-

•    Star schema        page 87

  • Hybrid schema
  • Pre-defined aggregate
  • All of the given options

The relation R will be in 2nd Normal Form if

  • It is in 1NF and each cell contains single value

•    It is in 1NF and each non key attribute is dependent upon entire primary key     page 44

  • It is in 1NF and non key attribute is dependent upon a single column of composite primary key
  • It is in 1NF and Primary key is composite
In ———–ested loop join of quadratic time complexity does not hurt the performance
    
 Typical OLTP environmentspage 22 
  • Data warehouse
  • DSS
  • OLAP

In Extract, Load, Transform(ELT) process, data transformation —————:

•    Takes place on the data warehouse server  page 147

  • Takes place on a separate transformation server
  • Depends on the nature of the source database
  • Does not take place

Node of a B-Tree is stored in memory block and traversing a B-Tree involves ————— page faults:

  • O(n log n)

•    O(log n)                     page 22

  • O(n)
  • O(n2)

As dimensions get less detailed (e.g. , year vs. day) cubes get ——————–

•    Smaller                    page 84

  • Larger
  • Partitioned
  • Merged

Which of the following is not a technique of “ Changed Data Capture” in currently used Modren

Source System?

  • Timestamps
  • Partitioning
  • Triggers

•    Dimensional Modeling page 150

The trade-offs of de-normalization is/are:

  • Storage
  • Performance
  • Ease-of-use

•    All of the given options    page 62

If actual data structure does not conform to documented formats then it is called:

•    Syntactically dirty data page 160

  • Semantically dirty data
  • Coverage anomaly
  • Extraction issue

“Header size is reduced, allowing more rows per back , thus reducing I/O” .The above statement is TRUE with respect to:

•    Vertical splitting             page 56

  • Horizontal splitting
  • Adding redundant column
  • None of the given options
  • Question: Break a teble into Multiple Tables based upon Comomn column values
  • Horizental Spliting
  • Vertical splitting
  • Adding redundant column
  • None of the given option

Which of the following is NOT an example of derived attribute?

  • Age
  • CGPA
  • Area of rectangle

•    Height        (Conceptual)

Which of the following is NOT an example of derived attribute?

  • Age
  • CGPA
  • Annual Salary

If a table is expected to have six columns but some or all of the records do not have six columns then it is example of:

•    Syntactically dirty data       page 160

  • Semantically dirty data
  • Coverage anomaly
  • Extraction issue

MDX by Microsoft is an example of ————————:

  • HOLAP
  • DOLAP
  • ROLAP

•    None of the given options        page 79

The growth of master files and magnetic tapes exploded around the mid- —————

  • 1950s

•    1960s               page 12

  • 1970s
  • 1980s

If one or more records in a relational table do not satisfy one or more integrity constraint , then the data:

  • Is syntactically dirty

•    Is semantically dirty        page 160

  • Has Coverage anomaly
  • Has extraction issue

OLAP is:

•    Analytical processing   page 69

  • Transaction processing
  • Additive processing
  • Active processing

One of the possible issues faced by web scrapping is that:

•    Web pages may contain junk data   page 141

  • Web pages do not contain multiple facts
  • Web pages do not contain multiple dimensions
  • Web pages does not support transformation

Which of the following is\are example of dimension:

•    Product      page 79

  • Region
  • Data
  • None of the given

An OLTP system is always good at ————————:

•    Evolving data  page 122

  • Keeping static data
  • Tracking past data
  • Maintaining historic data

In case of multiple sources for the same data element , we need to prioritize the source systems per element based, the process is called:

•    Ranking      page 143

  • Prioritization
  • Element selection
  • Measurement selection

One feature of Change Data Capture (CDC) is that:

  • It pre-calculates changed aggregates
  • It loads the transformed data in real time
  • It only processes the data has been changed

•    It can automate the transformation of extracted data  page 149

In —————— SQL generation in vastly simplified for front-end tools when the data is highly structure:

  • MOLAP

•    Star Schema                       page 107

  • Hybrid schema
  • Object oriented schema

Dirty data means:

  • Data cannot be aggregated
  • Data contains non-additive facts
  • Data does not fulfill dimensional modeling rules

•    Data does not conform to proper domain definitions   page 158

In Context of Change Data Capture (CDC) sometimes a ————- object can be used to store recently modified data:

  • Buffer table

•    Change table                 page 149

  • Checkmark table
  • Change control table

“Sometimes during data collection complete entities are missed”. This statement is an example of :

•    Missing tuple             page 161

  • Missing attribute
  • Missing aggregates
  • Semantically dirty data

Table collapsing technique is applied in case of:

•    One-by-one relation or many-to –many relation page 52

  • One-to-many relation
  • Many-to-many relation
  • None of the given option

Which of the following is an example of dimension?

  • Product
  • Region
  • Date

•    All of the given option page 78

Data warehouse stores ——————-:

  • Operational data

•    Historical data     page 24

  • Meta data
  • Log files data

The business process covered by ER diagrams:

•    Do not co-exist in time and space     page 109

  • Co-exist in time and space
  • Do not physically exist in real time context
  • None of the given options

The main goal of normalization is to eliminate ———–:

•    Data redundancy  page 41

  • Data sharing
  • Data security
  • Data consistency

Serious —- involves decomposing and resembling the data:

•    Data cleansing  page 168

  • Data transformation
  • Data loading
  • Data extraction

In the data warehouse environment the data is ————

•    Subject- oriented           page 69

  • Time- oriented
  • Both subject and time oriented
  • Neither time-oriented nor subject- oriented

For large record spaces and large number of records , the run time of the clustering algorithms:

•    Prohibitive             page 164

  • Static
    • •         Exponential
  • Numerical

————- can result in costly errors, such as , False frequency distributions and incorrect aggregates due to double counting:

•    Data duplication      page 165

  • Data reduction
  • Data anomaly
  • Data transformation

The degree to which values are present in the attributes that require them is known as –

———————:

•    Completeness         page 185

  • Uniqueness
  • Accessibility
  • Consistency

Time complexity of Key Creation process in basic Sorted Neighborhood (BSN) Method is

———————-:

  • O(n log n)
  • O(log n)

   O(n)             page 171

  • O(2n)

Which of the following is an example of slowly changing dimensions?

•    Inheritance      page 124

  • Aggregation
  • Association
  • Asset disposal

The ———— operator proves useful in more complex metrices applicable to the dimensions and accessibility:

•    Max              page 188

  • Min
  • Max and Min
  • None of the given

In OLAP , the typical write operation is ————- :

•    Bulk insertion    page 75

  • Single insertion
  • Sequential insertion
  • No insertion

The issue(s) of “ Adding redundant column” includes(s):

  • Increase in table size
  • Maintenance
  • Loss of information

•    All of the given option page 65

————– is applicable in Profitability analysis:

  • OLTP

•    Data warehouse         page 36,37

  • Information System(IS)
  • Management Information System(MIS)

The hardware (CPU) utilization in data warehouse environment is full or ———– :

  • Fixed
  • Partial

•    Not at all          page 24

  • Slow

Time variant is a characteristics of data warehouse which means:

•    Data loaded in data warehouse will be time stamped   page 20

  • Data can be loaded in data warehouse anytime
  • Data can be loaded in data warehouse only at a particular time
  • Data cannot be loaded in data warehouse with respect to time

In which class of aggregates AVERAGE function can be placed:

•    Algebraic           page 120

  • Distributed
  • Associative
  • Holistic

Considered the following Employee table and identify the column which causes that the table is not in first normal form(1NF): (Emp_ID, Emp_Name ,Emp_skills, Emp_Designation)

  • Emp_ID
  • Emp_Name

   Emp_skills           page 43(conceptual)

  • Emp_Designation

The application of data and information leads to ————-

  • Intelligence
  • Experience

•    Knowledge      page 11

  • Power

————— segregate data into separate partitions so that queries do not need to

examine all data in a table when WHERE clause filters specify only a subset of the partitions.

  • Pre-joining technique
  • Collapsing table technique

•    Horizontal splitting technique page 56

  • Vertical splitting technique

————-should not be present in a relation, so that it would be in second normal form (2NF).

•    Partial dependency         page 44 (conceptual)

  • Full functional dependency
  • Multivalued dependency
  • Transitive dependency

Records referring to the same entity are represented in different formulas in the different data sets or are represented erroneously. Thus duplicate records will appear in the merged database. This problem is known as————.

•    Merge/purge problem           page 168

  • Duplication problem
  • Redundant duplication problem
  • Redundant problem

The data perspective in OLTP system is operational, while that in data warehouse

is:

  • Fully normalized
  • Fully de-normalized
  • Fully summarized

•    Historical and detailed           page 30

Simple scalar transformation is a————–mapping from one set of values to another set of values using straightforward rules.

•    One-to-one       page 144

  • One-to-many
  • Many-to-many
  • Many-to-one

—————can be created in operational systems to keep tracks of recently updated records.

•    Triggers                   page 150

  • Timestamps
  • Partitioning
  • ELT

Development of data warehouse is hard because data sources are usually——–

  • Structured and homogeneous

•    Unstructured and heterogeneous     page 31

  • Structured and heterogeneous
  • Unstructured and homogeneous

In a decision support environment, the decision maker is interested in ————-.

  • Only limited organizational data

•    Big picture of organizational data    page 21

  • Only sale related data
  • Only customer related data

Information can answer question like “what”, “who” and “when” while knowledge can answer question like—————-.

  • Why
  • Where
  • Which

•    How                           page 11

OLTP implementations are fully————-.

•    Normalized                  page 69

  • Denormalized
  • Predictive

•    Additive

Which logical data extraction has significant performance impacts on the data warehouse server?

•    Incremental Extraction      page 133

  • Online Extraction
  • Offline Extraction
  • Legacy Vs OLTP

Consider the following Student table and identify the column which causes that the table is not in first normal form(1NF).

Student(Std_ID, Std_Name ,Std_CGPA ,Std_Hobbies)

  • •         Std_ID
  • Std_Name
  • Std_CGPA

•    Std_Hobbies       page 43(Conceptual)

Analytical processing uses —————

•    Multi-level aggregates     page 74

  • Record level aggregates
  • Table level aggregates
  • All of the given options

Which is not a class of anomalies in following?

•    Dirty anomalies  page 160

  • Syntactically dirty data
  • Semantically dirty data
  • Coverage anomalies

————- is a system of activities that assures conformance of product to pre-established requirements.

•    Quality assurance       page 183

  • Quality improvement
  • Quality Maintenance
  • Quality Establishment

Two interesting examples of quality dimensions that can make use of min operator are ——

•    Believability and appropriate amount of data    page 188

  • Believability and consistency
  • Believability and Redundancy
  • Reliability and appropriate amount of data

————– in database or data warehouse has no actual value; it only has potential

value.

•    Data            page 181

  • Entity
  • Flat tables
  • Data marts

In OLTP environment the selectivity is ———— and ———- in data warehouse environment.

•    High, Low       page 22

  • Low, High
  • High, Fixed
  • Fixed, Low

Which is not a/an characteristics of data quality?

•    Reliability           page 186

  • Uniqueness
  • Accessibility
  • Consistency

If a product meets formally defined “requirement specifications”, yet fails to be a quality product from the customer’s perspective , this means the requirements were ———–.

•    Defective         page 180

  • Unclear
  • Unrefined
  • Undefined

The relation R will be in 3rd Normal Form if:

  • It is in 2NF each cell contains single value
  • It is in 2NF and every non-key column is non-key transitively dependent upon its primary

key.           Page 46

  • It is in 1NF and each non key attribute is dependent upon a single column of composite primary key.
  • It is in 2NF and each non key attribute is dependent upon other non-key attribute.

Decision support system queries deal with number of columns ————

  • Having numeric values
  • In a single table
  • In a single view

•    Spanning across multiple tables       page 21

Normalization is used to reduce:

•    Reduces redundancy    page 41

  • Increases redundancy
  • Reduces joins

•    Reduces tables

The end user of data ware house are—————.

  • Programmers
  • Database developers
  • Data entry operator

•    Business executives        page 18 + 19

Which one are the characteristics of data warehouse queries?

  • Use primary key
  • High selectivity

•    Use multiple tables   page 30

  • Very low performance

Assume a company with a multi- million row customer table i.e. n rows. Checking for Referential Integrity (RI) using a naive approach would take —————— time.

•    O(n)            page 160

  • O(1)
  • O(log n)
  • None of the given

Web scrapping is a process of applying ————- techniques to the web

   Screen scrapping   page 146

  • Data scrapping
  • Text scrapping
  • Meta scrapping

Which is not an issue of ROLAP in the following?

•    Standard hierarchy of dimensions  page 92

  • Non-standard conventions
  • Maintenance
  • Aggregation

One of the fundamental purpose of de-normalization is to ——————— a number of physical tables which ultimately reduce the number of joins to answer a query.

  • Delete

•    Reduce  page 50

  • Increase
  • Decrease

———– is not the characteristic of data warehouse.

  • Time variant
  • Subject-oriented
  • Integrated

•    Volatile       page 69

Which is not a/an step of data cleansing procedure?

•    Aggregation              page 168

  • Elementizing
  • Standardizing
  • Verifying

Instance matching between different sources is then achieved by a standard ————-

on identifying attribute(s), if you are very, very, very lucky.

•    Equi-join      page 169

  • Inner join
  • Outer join
  • Fuller join

Ad-hoc access of data warehouse means:

  • That have predefined database access pattern

   That does not have predefined database access pattern        page 18

  • That could be accessed by any user
  • That could not be accessed by any user

In OLTP environment, the size of tables is relatively——————-

  • Large
  • Fixed
  • Moderate

   Small          page 22

————- is a/an measure of how current or up to date the data is:

•    Timeliness        page 185

  • Completeness
  • Accessibility
  • Consistency

The process of converting entity relationship model in to dimensional model of ———-

steps:

  • Two
  • Three

•    Four           page 109

  • Five

A ————- Is defined by a group of records that have similar characteristics (“behavior”) for p% of the fields in the data set, where p is a user- defined value(usually above 90).

•    Pattern        page 164

  • Cluster
  • Entity
  • Attribute

—————— is known as state of being only one of its kind or being without an equal or

parallel.

  • Completeness

•    Uniqueness             page 185

  • Accessibility
  • Consistency

Which of the following is not an example of fact?

•    Account no     page 74

  • Sales quantity
  • Per unit sales amount
  • Sales amount

——————is the degree to which data is accurately reflects the real world object that the data represents.

•    Intrinsic data quality   page 181

  • Realistic data quality
  • Strong data quality
  • Weak data quality

Which one among the following data warehouse stores data containing long period?

  • Telecommunication data warehouse
  • Financial data warehouse
  • Human resource data warehouse

   Insurance data warehouse  page 36

A ________ dimension is a collection of random transactional codes, flags and/text attributes that are unrelated to any particular dimension. The ______ dimension is simply a structure that provides a convenient place to store the ______ attributes.

  • Junk
    • Time
    • Parallel
    • None of these

Data Warehouse is about taking / collecting data from different ________ sources.

  • Harmonized
    • Identical
    • Homogeneous                         NOT CONFIRM
  • Heterogeneous

Taken jointly, the extract programs or naturally evolving systems formed a spider web, also known as

  • Distributed Systems Architecture
    • Legacy Systems Architecture
    • Online Systems Architecture
    • Intranet Systems Architecture

It is observed that every year the amount of data recorded in an organization

  • Doubles
    • Triples
  • Quartiles

The users of data warehouse are knowledge workers in other words they are _________ in the organization.

  • DWH Analyst
  • Decision maker
    • Database Administrator
    • Manager

Node of a B-Tree is stored in memory block and traversing a B-Tree involves ______ page faults.

  • O (n lg n)
  • O (log n) { O(log n) it’s the real answer}
    • O (n)
  • O (n2)

In _________ system, the contents change with time.

  • OLTP
    • ATM
    • DSS
    • OLAP

The growth of master files and magnetic tapes exploded around the mid- _______.

  • 1950s.
    • 1960s.
    • 1970s.
  • 1980s.

Relational databases allow you to navigate the data in ____________ that is appropriate using the primary, foreign key structure within the data model

  • Only One Direction
    • Any Direction
    • Two Direction
    • None of these

Naturally Evolving architecture occurred when an organization had a _______ approach to handling the whole process of hardware and software architecture.

  • Relaxed
    • Good
  • Not Relaxed
    • None

________ gives total view of an organization

  • OLAP
    • OLTP
  • Data Warehouse
    • Database

Suppose the amount of data recorded in an organization is doubled every year. This increase is

__________ .

  • Linear
    • Quadratic
  • Exponential
    • Logarithmic

Which people criticized Dimensional Modeling (DM) as being a data mart oriented approach?

Those that consider ER model as Data marts

Which of the following is not a CUBE operation?

ANSI SQL

If actual data structure does not confirm to documented formats then it is called:

Semantically dirty data

This technique can be used when a column from one table is frequently accessed in a large scale join in conjunction with

Adding redundant column

“Header size is reduced, allowing more rows per block, thus reducing I/O”. The above statement is TRUE with respect to:

Vertical splitting

Gives total view of an organization.

Data Warehouse

ROLAP provides access to information via a relational database using

ANSI standard SQL

The main reason(s) for the increase in cube size may be

All of the given

If each cell of Relation R contains a single value (no repeating values) then it is confirmed that

Relation R is in 3rd Normal Form but in 2nd Normal Form

In Extract, Load, Transformation (ELT) process, you don’t need to purchase extra devices to achieve parallelism because.

You already have parallel data warehouse servers

Which of the following is not a technique of “Changed Data Capture” in currently used Modern Source Systems?

Dimensional Modeling

Breaks a table into multiple tables based upon common column values

Horizontal splitting

Which of the following is NOT an example of derived attribute?

Height

Grain of a fact table means

The meaning of one fact table row

Which of the following is NOT an example of derived attribute?

Email Address

Which of the following is an example of Additive Facts?

Average

If a table is expected to have six columns but some or all of the records do not have six columns then it is example of:

Syntactically dirty data

Involves splitting a table by columns so that a group of columns is placed into the new table and the remaining

Vertical splitting

Normalization.

Reduces redundancy

De-normalization affects:

Database size and query performance

MDX by Microsoft is an example of                      .

None of the given options

The divide and conquer cube partitioning approach helps alleviate the          limitations of MOLAP implementation.

Scalability

Non uniform use of abbreviations, units and values refers to:

Syntactically dirty data

Allows download of “cube” structures to a desktop platform without the need for shared relational or cube server.

DOLAP

Node of a B-Tree is stored in memory block and traversing a B-Tree involves            page faults.

O (lg n)

The growth of master files and magnetic tapes exploded around the mid  

1960s.      

Leave a Reply

Your email address will not be published. Required fields are marked *