# TU Wien:Advanced Methods for Regression and Classification VU (Filzmoser)/Oral exam 2019-12-19

Jump to navigation
Jump to search

### 11:00[edit]

Participant 1:

- Trees: Regression trees, classification trees. General idea, criterion to minimized, for both cases. What measures of node impurity are available? How to avoid overfitting (pruning).
- SVM: criterion for linearly separable, non-linearly separable case. Kernel trick, kernel functions.
- GAM: for regression. How does the model look like, what functions mimimize criterion.

Participant 2:

- Mulitple regression model: Ordinary LS solution, how to arrive at it, what to do in near singularity of X^T X, R
- Ridge Regression, Lasso Regression, how does that look like, what is is different to OLS
- Spline regression: Criterion to minimize (with penalization of curvature), what functions do minimize this (natural cubic splines)

### 11:30[edit]

Participant 1:

- Define PCR and weighted least squares. Compare these two methods
- Define SVM for the linearly seperable and for the inseperable case.