Sunteți pe pagina 1din 11

Unit -I

SYLLABUS
Learning – Types of Machine Learning – Supervised Learning – The Brain and the Neuron – Design a
Learning System – Perspectives and Issues in Machine Learning – Concept Learning Task – Concept
Learning as Search –Finding a Maximally Specific Hypothesis – Version Spaces and the Candidate
Elimination Algorithm – Linear Discriminants – Perceptron – Linear Separability – Linear Regression.

PART – A
Q.No Questions Competence BT Level

1 What is machine learning?


Remember BTL1

2 Point out few examples of machine learning applications. Analyze BTL 4


3 Distinguish between supervised and unsupervised learning. Understand BTL 2
4 Define the types of machine learning. Remember BTL 1
5 Describe the steps to be followed for designing a Learning
Understand BTL 2
system for play checkers.
6 Discover one useful perspective on machine language. Apply BTL3
7 Explain the issues in machine learning. Analyze BTL 4
8 Explain concept learning. Evaluate BTL 5
9 Define Inductive Learning Hypothesis. Remember BTL 1
10 Assess the Enjoy sport learning task. Evaluate BTL 5
11 Define concept learning as search. Remember BTL 1
12 Develop candidate-elimination algorithm using version spaces. Create BTL 6
13 Define General-to-Specific Ordering of Hypotheses. Remember BTL 1
14 Identify the ways to represent The LIST-THEN-Eliminate
Remember BTL1
algorithm.
15 Examine find-s: finding a maximally specific hypothesis. Apply BTL 3
16 Discuss candidate-elimination algorithm. Understand BTL 2
17 Discover perceptron. Apply BTL 3
18 Develop the linear discriminants. Create BTL 6
19 Analyze version space. Analyze BTL 4
20 Describe the threelevels of analysis in information processing by
Understand BTL 2
brain.
PART-B (13 MARK )
1 (i)What is machine learning? Discuss about learning and machine
learning. Analyze BTL 4
(ii)Discuss the various types of machine learning
2 (i)Discuss in detail about Supervised learning. Understand BTL 2
(ii)Discuss about the Classification problem.
3 (i)Summarize the Issues in Machine Learning.
Evaluate BTL 5
(ii)Summarize the concept learning task.
4 (i)Describe the concept learning as search.
(ii)Describe General-to-Specific Ordering of Hypothesis for concept Remember BTL 1
learning as a search.
5 Explain find-s: finding a maximally specific hypothesis. Apply BTL 3
6 (i)Explain CANDIDATE-Elimination algorithm.
Apply BTL 3
(ii)Explain in detail about List-then –Eliminate algorithm.
7 With a Neat diagram, Explain about More Compact Representation Create BTL 6
for Version Spaces.
8 (i)Describe in detail about Linear Discriminants.
(ii)Discuss: Generalizing the Linear Model and Geometry of the Remember BTL 1
Linear Discriminant.
9 (i)Discuss about perceptron. Understand BTL 2
(ii)Discuss about the perceptron learning Algorithm with an example.
10 (i)Describe about the working of Brain and the Neuron
(ii)Explain about the Limitations of the McCulloch and Pitts Neuronal Remember BTL 1
Model
11 (i)Describe Find-S and explain the algorithm.
Remember BTL 1
(ii)Discuss about Inductive Bias.
12 Summarize in detail about Discrimination by Regression with an Understand BTL 2
example.
13 (i)Explain in detail about linear seperability.
Analyze BTL 4
(ii)Discuss Limitations of the McCulloch and Pitts Neuronal Model
14 Explain about Linear Regression with an example. Analyze BTL 4
PART - C (15 MARK )
1 Develop a program to learn to play. Checkers with goal of entering it
in the world checkers tournament.
(i)Choosing the Training Experience Create BTL 6
(ii)Choosing the Target Function
(iii)Choosing a Representation for the Target Function.
2 Explain the following for designing a learning system
(i)choosing a function approximation algorithm
(ii)estimating training values
(iii)adjusting the weights Analyze BTL 4
(iv)the final design.

3 (i)Describe in detail about Pair wise Separation of classes


Remember BTL 1
(ii) ExplainParametric Discrimination.
4 Describe about Vapnik-Chervonenkis (VC) Dimension.
Remember BTL 1

Unit -II
SYLLABUS
Multi-layer Perceptron – Going Forwards – Going Backwards: Back Propagation Error – Multi-layer
Perceptron in Practice – Example of using the MLP – Overview – Deriving Back-Propagation – Radial Basis
Functions and Splines – Concepts – RBF Network – Curse of Dimensionality – Interpolations and Basis
Functions – Support Vector Machines.

PART - A

Q.No Questions Competence BT Level

1 Analyze Back propagation Algorithm. Analyze BTL4

2 Write about Multilayer Perceptrons. Remember BTL1

3 List out different output activation functions. Remember BTL1

4 Discuss Local minima. Understand BTL2

5 What is testing, training and validation. Remember BTL1

6 Give difference between local and distributed representation. Understand BTL2

7 Write about receptive field. Remember BTL1

8 Name the two parts of MLP. Understand BTL2

9 Explain hybrid learning. Understand BTL2

10 Write the error function used for the perceptron. Remember BTL1

11 Illustrate curse of Dimensionality. Apply BTL 3


12 Describe the various steps of deriving Back Propagation. Apply BTL 3

13 What is support vector machine? Apply BTL 3

14 Explain the principles of gradient descent. Evaluate BTL 5

15 Define Biases. Remember BTL 1

Differentiate optimal separating hyperplane and soft margin


16 Analyze BTL 4
hyperplane.

17 Analyze RBF Network. Analyze BTL 4

18 Summarize Back propagation of error Evaluate BTL‐5

19 Justify radial basis function Create BTL‐6

20 How smoothing splines. Create BTL‐6

PART-B (13 MARK )

(i)Write about multilayer perceptron. Remember BTL 1


1 (ii)Draw the structure of MLP network with inputs and outputs
layer
2 (i) Define the Back propagation Rule. Remember BTL 1
(ii)Describe the training rule for output unit weights
3 (i)Write about Multilayer perceptron algorithm. Remember BTL 1
(ii)Describe how the MLP is designed to be a batch algorithm
4 (i)What is Local Minima? Remember BTL 1
(ii)Discuss in detail about picking up Momentum
Discuss the following
5 (i) Data preparation in MLP Understand BTL 2
(ii)Amount of training Data
(iii)Number of Hidden layers
6 (i)Describe the working behavior of support vector machine. Understand BTL 2
(ii)How to train test and validate the MLP.
7 (ii)Discuss Regression Problem of using MLP. Understand BTL 2
(ii) How Data compression is performed
8 (i)Illustrate Generalization of multivariate data. Apply BTL 3
(ii)Describe in detail about Curse of Dimensionality
9 Explain in detail about RBF network. Apply BTL 3

10 (i)Create a two class problem and describe support vector machine. Analyze BTL 4
(ii)Investigate how to choose smoothing parameter
11 (i)Explain the optimal separating hyperplane. Analyze BTL 4
(ii)Examine the soft margin hyperplane.
(i) Generate recurrent network and its equivalent unfolded
12 Analyze BTL 4
network.
(ii) Discuss about back propagation with time.
13 (i)Summarize about back propagation error. Evaluate BTL 5
(ii)Explain how the weights of the network are trained
(i)Justify your answer why we use Interpolation and Basis BTL 6
14 function. Create
(ii)Discuss in detail about Smoothing splines.
PART - C (15 MARK )
1 Write about the MLP as a Universal Approximator. Evaluate BTL5

(i)Create an example and explain about Principal Component


2 Analysis. Create BTL6
(ii)Describe Multiclass Discrimination.
3 Illustrate some examples of using MLP and the four types of Evaluate BTL5
problems that are generally solved using MLP.
4 Write a case study for the following
(i) Illustrate an example of using MLP
Create BTL 6
(ii) task involved
(iii) Input choice
(iv) Input encoding and output encoding.

Unit -III
SYLLABUS
Learning with Trees-Decision Trees-Constructing Decision Trees-Classification and Regression Trees-
Ensemble Learning-Boosting-Bagging-Different Ways to combine Classifiers-Probability and Learning –
Data into Probabilities-Basic Statistics-Gaussian Mixture Models-Nearest Neighbor Methods-Unsupervised
Learning – K means Algorithms – Vector Quantization – Self Organizing Feature Map.
.
PART - A

Q.No Questions Competence BT Level

1 Define Decision trees. Remember BTL1


2 What is decision node? Remember BTL1

3 What is regression tree? Remember BTL1

4 Illustrate an example for decision tree with diagram. Apply BTL3

5 Describe classification tree construction. Understand BTL2

6 List out the difference between pre and post pruning. Remember BTL1

7 Show the difference between Diversity and Accuracy. Apply BTL3


8 Define pruning. Remember BTL1

9 What is boosting? Understand BTL2

10 Classify the different ways to combine classifiers. Apply BTL3

11 Point out difference between regression tree and classification


Analyze BTL4
tree.

12 How to write weight update rule? Analyze BTL4

13 Explain what is Bayes optimal Classification Evaluate BTL5

14 Define bagging. Remember BTL1

15 What is No Free Lunch Theorem? Understand BTL2

16 Can you give few basic statistics concept? Evaluate BTL5

17 Perform k-means algorithm. Create BTL6

18 Compose kernel smoothers. Create BTL6

19 What vector quantization? Understand BTL2

20 Differentiate between k means and online k-means algorithm Analyze BTL4

PART-B (13 MARK )

1 (i)What is decision tree? Construct a decision tree. Remember BTL1


(ii)Write about univariate trees.
2 (i)Describe the Classification Trees. Remember BTL1
(ii)Draw and explain Entropy function for a class problem.
(i)Explain in detail about regression trees.
3 Analyze BTL4
(ii)How Regression tree smoothes for various values of Ѳr

4 Discuss the classification example for the decision tree. Understand BTL2

(i) Use decision tree to classify the students in a class based on


5 Apply BTL3
their academic.
(ii)Discuss about mixture of Experts Algorithm.
(i)What is Naïve Bayes Classifier?
6 (ii) Explain the following Basic statistics: Averages and Remember BTL1
Variance and Covariance.
7 (i)Illustrate Turning data into probabilities. Apply BTL3
(ii)Draw the histogram of feature values
(i) Write about Classification and Regression Trees (CART)
8 Remember BTL1
with examples.
(ii)Explain The Expectation-Maximization (Em) Algorithm.
Explain the following Statistics
9 Analyze BTL4
(i)Gaussian
(ii)Bias-variance Trade-off.
10 (i) Describe the Nearest Neighbor Algorithm. Understand BTL2
(ii)Write in detail about Nearest Neighbor Smoothing.

11 (i)Discuss the basic idea of ensemble learning. Understand BTL2


(ii)Write in detail about the various ensemble methods.
12 (i)Write about unsupervised learning Evaluate BTL5
(ii)Discuss the k-means algorithm in detail
13 Explain vector quantization in detail. Analyze BTL4

14 (i)Illustrate self organizing feature map Create BTL6


(ii)Illustrate SOM algorithm
PART - C (15 MARK )
1 (i)Create classification tree construction. Create BTL6
(ii)Describe regression tree algorithm.
2 (i)Write in detail about Boosting and AdaBoost Evaluate BTL 5
(ii)Discuss What is Bagging and Subagging
(i)Explain in detail about Gaussian Mixture Model.
3 Evaluate BTL 5
(ii)Write the General Expectation (EM) Algorithm.
(i) Design K-means algorithm and group the points (1, 0, 1), (1,
4 1, 0), (0, 0, 1) and (1, 1, 1) using K-means algorithm. Create BTL6
(ii)Perform normalization for the Neural network.

Unit -IV
SYLLABUS
Dimensionality Reduction-Linear Discriminant Analysis –Principal Component Analysis-Factor Analysis-
Independent component Analysis-Locally Linear Embedding –Isomap-Least Squares Optimization-
Evolutionary Learning-Genetic Algorithms-Genetic offspring Genetic Operators-using Genetic Algorithms-
Reinforcement Learning-Overview-Getting Lost –Example-Markov Decision Process.
PART - A

Q.No Questions Competence BT Level

1 Point out why dimensionality reduction is useful? Analyze BTL4


2 Define Factor Analysis or latent variables. Remember BTL1
3 Distinguish between within-class scatter and between-classes scatter. Understand BTL2
4 Define PCA. Remember BTL1
5 Describe what isomap is. Understand BTL2
6 Discover Locally Linear Embedding algorithm with k=12. Apply BTL3
7 Explain the three different ways to do dimensionality reduction. Analyze BTL4
8 Explain what Least Squares Optimization is. Evaluate BTL5
9 Difference action and state space. Remember BTL1
10 Write what is Punctuated Equilibrium? Evaluate BTL5
11 How reinforcement learner experience and the corresponding action. Remember BTL1
12 Express the basic tasks that need to be performed for GA. Create BTL6
13 Define TD-Gammon Remember BTL1
14 Identify how reinforcement learning maps states to action. Remember BTL1
15 Examine Genetic Programming. Apply BTL3
16 Discuss about reward function. Understand BTL2
17 Discover Markov decision processes. Apply BTL3
18 Differentiate Sarsa and Q-learning. Create BTL6
19 Analyze POMDPs. Analyze BTL4
20 Describe about values. Understand BTL2
PART-B (13 MARK )
(i)Write the three method s of dimensionality reduction.
1 (ii)Discuss how to choose the right features with an example and Analyze BTL4
perform the various analysis.
2 (i)Discuss in detail LDA. Understand BTL2
(ii)Discuss how to measure the dataset by various method.
3 (i)Write in detail factor analysis Evaluate BTL5
(ii)Write the output of using factor analysis on the iris data set.
4 (i) Describe Locally Linear Embedding algorithm. Remember BTL1
(ii) Describe LLE algorithm
5 (i) Explain in detail about isomap Apply BTL3
(ii) Discover the result of applying isomap to the iris dataset
6 Explain Least Squares Optimization. Create BTL6

7 (i)Describe Evolutionary Learning Remember BTL1


(ii) Describe about population and Parent selection
8 (i)Describe in detail about Generating Offspring Genetic Operators Understand BTL2
(ii)Discuss the Basic Genetic Algorithm
9 (i)Identify the difference between Sarsa and Q-learning Remember BTL1
(ii)Discuss an example for the reinforcement learning
10 (i)Describe Knapsack problem for GA Remember BTL1
(ii)Describe the for peaks Problem for GA
11 (i)Write in detail about reinforcement learning Create BTL6
(ii)Illustrate the reinforcement learning cycle
(i)Write about the Markov Property.
12 (ii) Generalize how principal component analysis is carried out to Create BTL 6
reduce dimensionality of data sets.
13 (i)Write in detail about values Remember BTL 1
(ii)Show an example of environment
Describe in detail about
14 (i)Getting Lost Create BTL 6
(ii) Discounting
PART - C (15 MARK )
1 Generalize PCA algorithm and discuss Kernel PCA. Create BTL 6
Write in detail about Np complete problem and show the string
2 Create BTL 6
representation.

3 Choose two destination with different routes connecting them. Apply Analyze BTL 4
genetic algorithm to find the optional path based on distance.
(i)Explain about reinforcement learning.
4 Evaluate BTL 5
(ii) Explain about isomap ? Give its significance in machine learning.
Unit -V
SYLLABUS
Markov Chain Monte Carlo Methods-Sampling –Proposal Distribution-Markov Chain Monte Carlo-Graphical
Models –Bayesian Networks-Markov Random Fields-Hidden Markov Models-Tracking Methods.

PART - A

Q.No Questions Competence BT Level

1 Point the algorithm that produce pseudo-random numbers Analyze BTL4


2 Define MCMC Remember BTL1
3 Distinguish random numbers and Gaussian Random numbers Understand BTL2
4 Define Sampling Remember BTL1
5 Describe Markov Chains Understand BTL2
6 Discover the proposal distribution method Apply BTL3
7 Explain the Various mcmc methods Analyze BTL4
8 Explain graphical models Evaluate BTL5
Differentiate between sampling –importance sampling and resampling
9 Remember BTL1
algorithm
10 Write Gibbs sampler Evaluate BTL5
11 How gibbs sampler forms the basis for software package Remember BTL1
12 Express Bayesian Belief network Create BTL6
13 Difference between: Directed and Undirected graphs Remember BTL1
14 Identify polytrees in a graph Remember BTL1
15 Examine approximate inference. Apply BTL3
16 Discuss Making Bayesian networks Understand BTL2
17 Discover HMM Apply BTL3
18 Differentiate top –down and bottom-up inference Create BTL6
19 Analyze HMM Baum-Welch(forward –Backward )algorithm Analyze BTL4
20 Describe the two Tracking methods Understand BTL2
PART-B (13 MARK )
1 (i)Write about sampling. Analyze BTL4
(ii)Discuss the sampling methods in generation of random numbers

2 (i)Discuss Gaussian Random numbers Understand BTL2


(ii)Describe Box-Muller scheme
3 (i) Write the Rejection Sampling Algorithm Evaluate BTL5
(ii)Show the histogram of a mixture of two Gaussians
4 (i) Describe Sampling –importance resampling algorithm Remember BTL1
(ii) Describe Gibbs Sampling
5 (i) Explain in detail about MCMC Apply BTL3
(ii) Discover Simulated Annealing by a distribution
(i)Structure two graphical models and show the various relationship
6 between the nodes. Create BTL6
ii). Explain about conditional table
7 (i).Describe variable elimination algorithm Remember BTL1
(ii)Describe the Approximate Inference
8 (i)Describe in detail about Bayesian Network. Understand BTL2
(ii)Explain with an example
9 Identify the structure and conditional probability Remember BTL1
(i)Describe in detail Markov Random fields
10 Remember BTL1
(ii)Write the Markov Random Field Image Denoising Algorithm

11 (i)Write in detail forward algorithm Create BTL6


(ii)Discuss HMM forward algorithm
12 Write a HMM Baum-Welch(forward –Backward )algorithm Create BTL 6

13 (i)Write about Tracking methods Remember BTL 1


(ii) Identify hidden Markov models in detail.
14 (i)Describe in detail about Kalman Filter Algorithm Create BTL 6
(ii)Discuss about The particle Filter
PART - C (15 MARK )
1 Generalize the proposal Distribution. Create BTL 6

2 (i)Write about Markov chains. Create BTL 6


(ii)How Metropolis-Hastings algorithm rejects or accepts a sample
3 Explain about Markov Chain Monte Carlo Methods. Analyze BTL 4
4 Measure the decoding problem by viterbi algorithm Evaluate BTL 5

S-ar putea să vă placă și