Sie sind auf Seite 1von 4

DATA MINING AND DATA WAREHOUSING

[As per Choice Based Credit System (CBCS) scheme]


(Effective from the academic year 2017 - 2018)
SEMESTER – VI

Subject Code : 17CS651 IA Marks :40


Number of Lecture Hours/Week : 3 Exam Marks :60
Total Number of Lecture Hours : 40 Exam Hours :03

Module 1
1. Explain with diagram, a three-tier data warehouse architecture. (10 Marks Jan 2019,July
2019)
2. Explain OLAP operations with examples. (10 Marks Jan 2019,July 2019,2018)
3. Define:1) Dimensions’ 11) Measures 111) Fact tables. (03 Marks Jan 2019)
4. What is a Data warehouse and what are its four key feature? (4 Marks July 2019)
5. Define data warehouse. With neat diagram, explain implementation of operational data
Store. (10 Marks July 2018)
6. a. What is data warehouse? Differentiate between ODS and data warehouse. (06 Marks
Jan 2019)
b. What is metadata in data warehouse? What it contains? (03 Marks Jan 2019)
7. Compare OLTP and OLAP Systems. (06 Marks July 2019,2018,jan 2018)
8. Explain the following terms with examples: 1) Snow Flake schema ii) Fact Constellation
schema iii) Star schema (9 Marks july 2019)
9. Give the definition of data warehousing, Discuss the need for data warehousing.(6 Marks
ja 2018)
10. Discuss the characteristics of operational data store with its design and implementation
issues.(10 Marks Jan 2018)
11. Describe the operations of Data Cube.( 10 Marks jan 2018)
12. With suitable example. explain star schema for multidimensional database. (10 Marks
july 2018)
13. Explain cube operations with suitable examples. (10 Marks july 2018)
14. Present five major characteristics from Codd's rule.(5 Marks jan 2018)
Module 2
1. Describe ROLAP, MOLAP, HOLAP. (10 Marks jan 2019)
2. Explain indexing OLAP data: Bitmap index and join index with example. (10 Marks jan
2019)
3. What is Data Mining? With a neat diagram, Explain the KDD process in Data Mining.
(10 Marks jul 2019)
4. Whatis data mining? Briefly explain the motivating challenges. (06 Marks jan 2019)
5. Explain ROLAP versus MOLAP. (06 Marks jan 2019,2018).
6.
7. Explain KDD process. Highlight the importance of four core data mining tasks. (10
Marks July 2018)
8. Define Data preprocessing. Mention the steps involved in it. Explain any 2 steps in detail.
(10 Marks June 2019)
9. Explain data preprocessing steps. (10 Marks jan 2019)
10. Describe the various type of attributes and data sets. (10 Marks June 2019)
11. For the following vectors X and Y. Calculate cosine similarity where X = {3 2 0 5 0 0 0
2 0 0}, Y{1 0 0 0 0 0 0 1 0 2} (10 Marks June 2019)

Module 3

1. Explain frequent itemset generation of the Apriori algorithm. (08 Marks jan 2019, jul
2019)
2. Explain rule generation in Apriori algorithm. (08 Marks jan 2018)
3. Explain the following terms with example : i) Rule generations; ii) Computational
complexity.(6 Marks Jul 2019)
4. With suitable example illustrate the Apriort principle. (10 Marks jul 2018).
5. What is Frequent Itemset Generation? Explain Frequent Itemset Generation using Apriori
principle.(10 marks jan 2018)
6. Generate frequent itemset for the given data with support - 50%.( 5 Marks jul 2019)
TID 100 200 300 400
item {1,3,4} {2,3,5} {1,2,3,5} {2,5}
s

7. For the following transaction data set : i) Construct an FP tree ii) Generate the list of
frequent itemset. Ordered by their corresponding suffixes. ( 9 Marks jul 2019)
TID 1 2 3 4 5 6 7 8 9 10

Item {a,b {b,c,d {a,c,d,e {a,d,e {a,b,c {a,b,c,d {a {a,b,c {a,b,d {b,c,e
s } } } } } } } } } }

8. Briefly explain the candidate generation procedure F k−1 × F k−1 Merging strategy.(7
Marks jul 2019)
9. Explain alternative methods for generating frequent itemsets. (08 Marks jan 2018)
10. Explain briefly FP-growth algorithm. (08 Marks jan 2018)
11. Explain frequent itemset generation in FP-growth algorithm. (10 Marks jul 2018)
12. Given the following set of transactions in market basket model. Build a frequency
pattern(FP tree) show each transaction separately. (10 Marks jan 2018)
Transaction Itemset
1 Milk,Bread,Cookies,Juice
2 Milk, Juice
3 Milk,eggs
4 Bread, Cookies,
5 Juice,eggs
6 Bread,eggs

Module 4

1. Explain how decision tree induction algorithm works. Give example. ( 8 Marks jul
2019)
2. List and explain the different characteristics of decision tree induction. (8 Marks jul
2019)
3. Describe the nearest neighbour classification techniques. (8 Marks jul 2019)
4. Write a note on Bayesian classifier. (8 Marks jul 2019)
5. With neat block diagram, explain general approach to solve classification problem.(08
Marks jan 2019)
6. Explain how to build a decision tree using Hunt’s algorithm. (08 Marks jan 2019)
7. Explain rule based classifiers with illustration. (08 Marks jan 2019)
8. Explain K-nearest neighbor classification algorithm with example. (08 Marks jan 2019)
9. Explain how predictive accuracy of classification method is estimated.(10 marks Jan
2018)
10. What is rule-based classifier’? With example explain how a rule based classifier works.
(10 Marks jul 2018)
11. What are Bayesian classifiers? State Baye’s theorem and explain how Baye’s theorem ts
used in the Naive Bayestan classifier with example. (10 Marks jul 2018, jan 2018)

Module 5
1. What is Cluster analysis? Describe the different types of clustering. (8 Marks jun
2019,jan 2019,jul 2018, jan 2018)
2. Explain the following terms :i) K - means clustering li) Graph based clustering (8 Marks
jun 2019,jul 2018)
3. What are the basic approaches used for generating agglomerative hierarchical
clustering? (8 Marks jun 2019,jan 2019)
4. Explain the following terms :i) K - means clustering li) Divisive heirarchial clustering (8
Marks jan 2018)
5. Explain D B Scan algorithm, with example. (8 Marks jun 2019,jan 2019)
6. Briefly explain BIRCH scalable clustering algorithm. (08 Marks jan 2019)

Das könnte Ihnen auch gefallen