File Download
Supplementary
-
Citations:
- Appears in Collections:
postgraduate thesis: Improving discrete AdaBoost for classification by randomization methods
Title | Improving discrete AdaBoost for classification by randomization methods |
---|---|
Authors | |
Issue Date | 2016 |
Publisher | The University of Hong Kong (Pokfulam, Hong Kong) |
Citation | Dong, F. [董凤娇]. (2016). Improving discrete AdaBoost for classification by randomization methods. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. Retrieved from http://dx.doi.org/10.5353/th_b5760964 |
Abstract | Adaboost, a typical boosting method for classification, performs well in classification problems. Many researchers have applied different types of randomization techniques to Adaboost for further improving the efficiency of classification. However, these methods of randomization seldom aim at the chance mechanism underlying the training data itself, especially at the response level. We propose a new modified Adaboost procedure which takes into account the chance mechanism. Three different methods are investigated for estimating the conditional probabilities of class labels given feature covariates, based on which the class labels are randomized within the training dataset. The first method, which we term quantile calibration, makes use of a reweighting scheme to find a reliable interval containing the conditional class probability. The second method applies Bootstrap Aggregating to obtain an equal weight ensemble vote for each class label. The third method exploits a well-known connection between the score function of AdaBoost and class probabilities under an additive logistic regression setup. Empirical results show that our new procedure successfully alleviates the overfitting problem, and in many cases improves the classification performance of Adaboost as well. |
Degree | Master of Philosophy |
Subject | Boosting (Algorithms) |
Dept/Program | Statistics and Actuarial Science |
Persistent Identifier | http://hdl.handle.net/10722/226781 |
HKU Library Item ID | b5760964 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Dong, Fengjiao | - |
dc.contributor.author | 董凤娇 | - |
dc.date.accessioned | 2016-06-30T04:24:09Z | - |
dc.date.available | 2016-06-30T04:24:09Z | - |
dc.date.issued | 2016 | - |
dc.identifier.citation | Dong, F. [董凤娇]. (2016). Improving discrete AdaBoost for classification by randomization methods. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. Retrieved from http://dx.doi.org/10.5353/th_b5760964 | - |
dc.identifier.uri | http://hdl.handle.net/10722/226781 | - |
dc.description.abstract | Adaboost, a typical boosting method for classification, performs well in classification problems. Many researchers have applied different types of randomization techniques to Adaboost for further improving the efficiency of classification. However, these methods of randomization seldom aim at the chance mechanism underlying the training data itself, especially at the response level. We propose a new modified Adaboost procedure which takes into account the chance mechanism. Three different methods are investigated for estimating the conditional probabilities of class labels given feature covariates, based on which the class labels are randomized within the training dataset. The first method, which we term quantile calibration, makes use of a reweighting scheme to find a reliable interval containing the conditional class probability. The second method applies Bootstrap Aggregating to obtain an equal weight ensemble vote for each class label. The third method exploits a well-known connection between the score function of AdaBoost and class probabilities under an additive logistic regression setup. Empirical results show that our new procedure successfully alleviates the overfitting problem, and in many cases improves the classification performance of Adaboost as well. | - |
dc.language | eng | - |
dc.publisher | The University of Hong Kong (Pokfulam, Hong Kong) | - |
dc.relation.ispartof | HKU Theses Online (HKUTO) | - |
dc.rights | The author retains all proprietary rights, (such as patent rights) and the right to use in future works. | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.subject.lcsh | Boosting (Algorithms) | - |
dc.title | Improving discrete AdaBoost for classification by randomization methods | - |
dc.type | PG_Thesis | - |
dc.identifier.hkul | b5760964 | - |
dc.description.thesisname | Master of Philosophy | - |
dc.description.thesislevel | Master | - |
dc.description.thesisdiscipline | Statistics and Actuarial Science | - |
dc.description.nature | published_or_final_version | - |
dc.identifier.doi | 10.5353/th_b5760964 | - |
dc.identifier.mmsid | 991019898389703414 | - |