Sie sind auf Seite 1von 3

Date of showing evaluated answer books: ..

/12/2018

No. of Printed Pages: 01 Subject Code

I C S 7 0 3

B.TECH ODD SEMESTER EXAMINATION, 2018-19


NAME OF SUBJECT: EL-I (DATA WAREHOUSING & DATA MINING)

Time: 3 Hrs. Max. Marks: 100

Note: 1. Attempt all questions.


2. All questions carry marks as shown against them.
1. (a) What is on-line transaction processing (OLTP)? Describe the evolution of OLTP. (10)
What are the critical features of OLTP systems?
(b) What is data integration and transformation? Explain data cube aggregation in brief. (05)
(c) What is a concept hierarchy? Describe the OLAP operations in the Multidimensional (05)
data model.
(OR)

What is the difference between OLTP and OLAP systems?


2. (a) What is data cleaning? Discuss various ways of handling missing values during data (05)
cleaning?
(b) What are the various components in a multi-tiered Data warehouse architecture? (05)
(c) What are Bayesian classifiers? With an example, describe how to predict a class label (05)
using naive Bayesian classification.

(d) How does the data warehousing and data mining work together. (05)

(OR)

Discuss the process of K-Mean clustering. Also outline major drawbacks of K- Mean
clustering technique.

3 (a) What is market-basket analysis? What type of mining is possible from this data? (05)
What are the parameters used in the process?
(b) State the Apriori Property. Generate large itemsets and association rules using (10)
Apriori algorithm on the following data set with minimum support value and
minimum confidence value set as 50% and 75% respectively.
TID Items Purchased
T101 Cheese, Milk, Cookies
T102 Butter, Milk, Bread
T103 Cheese, Butter, Milk, Bread
T104 Butter, Bread
(c) Explain various methods to improve the efficiency of the Apriori- based (05)
mining.
(OR)

How K-Mean clustering method differs from K-Medoid clustering method?


4. (a) How does k-medoids algorithm cluster in data? What are its advantages?. (05)

1
(b) What is supervised learning? Using the given table, show how the ROOT splitting (05)
attribute is selected using InfoGain measure in the overall process of decision tree
induction.

SL. No. Attributes Class

Outlook Temperature Humidity Windy

1 Sunny Hot High False N

2 Sunny Hot High True N

3 Overcast Hot High False P

4 Rain Mild High False P

5 Rain Cool Normal False P

6 Rain Cool Normal True N

7 Overcast Cool Normal True P

8 Sunny Mild High False N

9 Sunny Cool Normal False P

10 Rain Mild Normal False P

11 Sunny Mild Normal True P

12 Overcast Mild High True P

13 Overcast Hot Normal False P

14 Rain Mild High True N

(c) What is Hierarchical clustering? Describe any one Hierarchical clustering algorithm. (05)
(d) Compare association and classification. Briefly explain associative classification (05)
with suitable example.

(OR)

What is meant by “clustering”? Explain why clustering is called unsupervised


learning. Mention any two applications of clustering.
5. (a) Cluster following points in three clusters. Take initially A1, B1 and C1 as Centre (10)
points Use k-means algorithm and show the final clusters formed (Use Euclidean
distance.)
A1(2,10), A2(2,5), A3(8,4), B1(5,8), B2(7,5), B3(6,4), C1(1,2), C2 (4,9).

2
(b) Explain about FP-growth. Generate fp-tree using a suitable example with minimum (10)
support count is 3.

(OR)

What is Cuboid? Explain various OLAP operations on data cube with suitable
example.

Das könnte Ihnen auch gefallen