I run the cross-validation experiment for a given data set, and tried two different approaches: one is based on SVM, another is based on SVM plus Adaboost. But the confusion matrix for two experiments are exactly the same. I am confused on how to explain this kind of result. Adaboost is supposed to start with a weak classifier, but how to determine whether a classifier is weak?
Adaboost – How to Determine Whether a Classifier Like Adaboost is Weak
boostingclassificationdata miningmachine learningsvm
Related Question
- Solved – Lower classification rate than expected by chance
- Solved – How to derive confidence intervals from the confusion matrix for a classifier
- Quadratic Kernel – How Does a Quadratic Kernel Look Like in SVM?
- Solved – How to get accuracy, confusion matrix of binary SVM classifier equivalent to multiclass classification
- Solved – Boosting: In the function gbm() from library gbm and understanding the cv.folds implementation
- Machine-Learning – Proof of Upper Bound on Leave-One-Out Cross-Validation Error of Linear SVM
Best Answer
I can answer the second question. A weak classifier for AdaBoost means a classifier with low Variance (and consequently high bias) in Bias-Variance decomposition. The reason is, that combining classifiers increases variance, so when you start with high variance classifiers, you're very likely to overfit.