The questions on the final will be drawn from the entire semester. For examples of questions covering the material presented before midterms 1 and 2 check out these links:

There will also be questions related to the papers presented by the students in class. At most 80 points of the final will relate to these papers. Below are some potential questions based on the papers presented by students. The questions regarding those papers will be drawn from THIS list:

1. How does the Sequential Minimal Optimization (SMO) method for training a
   Support Vector Machine work?  Make sure to discuss how the notion of the
   support vectors figure into this, what is meant by the term "smallest
   possible optimization problems" (and what such a problem represents), how
   the method chooses from amongst these problems (and how that choice affects
   the solution method).  You should also compare the SMO method to the
   Chunking method for training SVMs.

2. How does the method for Support Vector Clustering introduced by Ben-Hur
   et al. work?  What does the support vector machine attempt to do in this
   method?  Also, how is the graph produced from the resulting machine and
   what part does the connected components algorithm play in producing the
   clusters?

3. Kondor and Jebara introduced a kernel function that would be useful in
   what types of situations?  Give an example of such a situation.  What are
   the advantages of this kernel over other kernels for comparing data?

4. Explain the basic idea behind Principal Component Analysis (PCA)?  In what
   types of domains would be PCA be very effective?  What type of problem
   did Suykens et al. propose to solve to perform PCA and what mechanism
   do they use in solving this problem?

5. Explain how Valentini and Dietterich proposed to control the Bias-Variance
   tradeoff in SVMs and how they proposed to use the resulting SVMs in an
   ensemble learning method.

6. In the Jaakola et al. paper "A discriminative framework for detecting
   remote protein homologies" explain what is meant by the term remote
   homology?  In this paper Jaakola et al. proposed to use compare generative
   models - what type of model and how would this models be compared (what
   mechanism would be used in the comparison and what would be compared)?

7. Vinokourov et al. proposed a method that would allow (amongst other things)
   for cross language queries (a query in one language that could find a 
   relevant document in another language).  Explain the basics of their
   method and how such a query would be possible.

8. In the Bayes meets Bellman paper, Engel et al. proposed to learn a policy
   function as a Gaussian process.  What functions did they propose to learn
   to produce a policy (and what assumptions are built into this approach)?  
   What are the advantages of this approach?  What do they mean by the term
   on-line sparsification and how is it important to their approach?

9. What is meant by the term "model" in Wang and Dietterich's "Model-based
   policy gradient reinforcement learning"?  How is this model learned and
   how is it used in their approach?  What are the advantages and disadvantages
   of this approach with respect to other reinforcement learning methods?

10. What is the difference between an Markov Decision Process (MDP) and
   a Semi Markov Decision Process (SMDP)?  Explain how this relates to the
   use of temporally extended actions such as those introduced by Options
   (Sutton, Precup and Singh)?

11. What is an Option as introduced by Sutton, Precup and Singh?  How are
   options used in a reinforcement learner?  How can an option help speedup
   learning?

12. What is a Hierarchical Abstract Machine (HAM) as introduced by Parr and
    Russell? Who creates HAMs?  How does a reinforcement learner work in
    a system involving HAMs?  What is a choice point?