The figure below shows the available properties of AdaBoost:-
Field | Description | Remark | |
Run | It allows you to run the node. | - | |
Explore | It allows you to explore the successfully executed node. | - | |
Vertical Ellipses | The available options are
| - | |
Task Name | It is the name of the task selected on the workbook canvas. | You can click the text field to edit or modify the task name as required. | |
Dependant Variable | It allows you to select the dependent variable | You can choose only one variable. It should be of a Categorical type. | |
Independent Variable | It allows you to select the independent variable. | You can select more than one variable. | |
Advanced | Learning Rate | It allows you to change the learning rate accordingly | When the learning rate is higher, it leads to a greater contribution of each classifier. |
Number of Estimators | It allows you to select the number of estimators. |
| |
Algorithm | It allows you to select between the two given options | The options are SAMME and SAMME.R | |
Random State | It allows you to enter the value of the random state. | Enter only numerical value. | |
Dimensionality Reduction | It allows you to select the dimensionality reduction method. |
|
In the example provided below, the Superstore dataset is used to apply AdaBoost. The independent variables considered are City, Sales, and Profit, while the dependent variable selected is Category.
After using the AdaBoost algorithm, the following results are displayed.
The result page displays the following sections.
Section 1 – Key Performance Index (KPI)
The categorical variable's different options are displayed in the top right corner. Here Furniture variable is displayed. The first option appears as the default selected option.
Field | Description | Remark |
Accuracy | Accuracy is the ratio of the total number of correct predictions made by the model to the total number of predictions made. Accuracy = (TP + TN) / (TP + TN + FP + FN) | The Accuracy is 0.6559. |
F-Score | F-score is a measure of the accuracy of a test. |
|
Precision | Precision is the ratio of the True positive to the sum of the True positive and False Positive. It represents positive predicted values by the model. | Here Precision is 0.6232. |
Sensitivity | It measures the test's ability to identify positive results. |
|
Specificity | It gives the ratio of the correctly classified negative samples to the total number of negative samples: Specificity = TN / (TN + FP) |
|
Section 2 – Confusion Matrix
A confusion matrix is a summarized table used to assess the performance of a classification model. The number of correct and incorrect predictions is summarized with count values and broken down by each class.
Following is the confusion matrix for the specified categorical variable. It contains predicted values and actual values for the Category.
Section 3 – ROC chart
The Receiver Operating Characteristic (ROC) Chart is given below. The ROC curve is a probability curve that measures the performance of a classification model at various threshold settings.
Section 4 – Lift Chart
The Lift Chart obtained is given below. A lift is the measure of the effectiveness of a model. It is the ratio of the percentage gain to the percentage of random expectation at a given decile level. It is the ratio of the result obtained with a predictive model to that obtained without it.