"AdaBoost, or adaptive boosting is an ensemble learning method that uses iteration to learn from the errors of weak classifiers to develop strong classifiers. It is based on the concept that a single weak classifier may not be able to accurately predict the class of an object (slightly better than a guess), but when multiple weak classifiers are combined, a strong classifier can be created. \n",
"* AdaBoost is not a model, it is applied on top other types of classifiers, e.g. logistic regression or decision trees, to learn from their limitations and suggest an improved model. \n",
"* AdaBoost is sequential. Firstly, a model is built based on training data. This model is evaluated, then another model is trained to specifically correct the errors of the first model by increasing the weights of the misclassifed examples and decreasing the weights of the correctly classified examples. \n",
"* This process is repeated iteratively until the model cannot be improved or a predefined limit of iterations is reached.\n",
"* Error is reduced very rapidly, meaning AdaBoost has a high convergence rate\n",
"* Predictions are made by calculating the weighted majority vote or weighted average of the weak classifiers"
"C:\\Users\\Andreas Shepley\\Anaconda3\\lib\\site-packages\\sklearn\\model_selection\\_search.py:841: DeprecationWarning: The default of the `iid` parameter will change from True to False in version 0.22 and will be removed in 0.24. This will change numeric results when test-set sizes are unequal.\n",
"plt.title(\"Visualizing Important Features\")\n",
"plt.legend()\n",
"plt.show()"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"### Exercise 1: \n",
"Use Adaboost to predict the incidence of diabetes amongst heart patients represented by this dataset."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": []
}
],
"metadata": {
"kernelspec": {
"display_name": "Python 3",
"language": "python",
"name": "python3"
},
"language_info": {
"codemirror_mode": {
"name": "ipython",
"version": 3
},
"file_extension": ".py",
"mimetype": "text/x-python",
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.7.1"
}
},
"nbformat": 4,
"nbformat_minor": 2
}
%% Cell type:markdown id: tags:
## AdaBoost
%% Cell type:markdown id: tags:
AdaBoost, or adaptive boosting is an ensemble learning method that uses iteration to learn from the errors of weak classifiers to develop strong classifiers. It is based on the concept that a single weak classifier may not be able to accurately predict the class of an object (slightly better than a guess), but when multiple weak classifiers are combined, a strong classifier can be created.
* AdaBoost is not a model, it is applied on top other types of classifiers, e.g. logistic regression or decision trees, to learn from their limitations and suggest an improved model.
* AdaBoost is sequential. Firstly, a model is built based on training data. This model is evaluated, then another model is trained to specifically correct the errors of the first model by increasing the weights of the misclassifed examples and decreasing the weights of the correctly classified examples.
* This process is repeated iteratively until the model cannot be improved or a predefined limit of iterations is reached.
* Error is reduced very rapidly, meaning AdaBoost has a high convergence rate
* Predictions are made by calculating the weighted majority vote or weighted average of the weak classifiers
C:\Users\Andreas Shepley\Anaconda3\lib\site-packages\sklearn\model_selection\_search.py:841: DeprecationWarning: The default of the `iid` parameter will change from True to False in version 0.22 and will be removed in 0.24. This will change numeric results when test-set sizes are unequal.