TU Wien:Advanced Methods for Regression and Classification VU (Filzmoser)/Oral exam 2019-12-19

From VoWi
Jump to navigation Jump to search

11:00[edit]

Participant 1:

  • Trees: Regression trees, classification trees. General idea, criterion to minimized, for both cases. What measures of node impurity are available? How to avoid overfitting (pruning).
  • SVM: criterion for linearly separable, non-linearly separable case. Kernel trick, kernel functions.
  • GAM: for regression. How does the model look like, what functions mimimize criterion.

Participant 2:

  • Mulitple regression model: Ordinary LS solution, how to arrive at it, what to do in near singularity of X^T X, R
  • Ridge Regression, Lasso Regression, how does that look like, what is is different to OLS
  • Spline regression: Criterion to minimize (with penalization of curvature), what functions do minimize this (natural cubic splines)

11:30[edit]

Participant 1:

  • Define PCR and weighted least squares. Compare these two methods
  • Define SVM for the linearly seperable and for the inseperable case.