File Download
Supplementary

postgraduate thesis: Supervised machine learning with Delaunay triangulation learner and collaborative gradient boosting

TitleSupervised machine learning with Delaunay triangulation learner and collaborative gradient boosting
Authors
Advisors
Advisor(s):Yin, G
Issue Date2021
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Liu, Y. [刘业鸿]. (2021). Supervised machine learning with Delaunay triangulation learner and collaborative gradient boosting. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractIn this thesis, we discuss some statistical machine learning methods and their ensemble learning approachs with theories and applications. Firstly, we propose a new piecewise linear learner, called the Delaunay triangulation learner (DTL). Based on the data samples in a p-dimensional feature space, the Delaunay triangulation algorithm provides a unique triangulation of the space. The triangulation separates the convex hull of the samples into a series of disjoint p-simplices, where the samples are the vertices of the p-simplices. The DTL is constructed by fitting the responses through linear interpolation functions on each of the Delaunay simplices, and thus it approximates the whole functional by a piecewise linear function. In its ensemble learning approaches, we propose the bagging DTLs, random crystal and the boosting DTL, where the DTLs are constructed on the subspaces of the features, and the feature interactions are captured by Delaunay triangle meshes. Extensive numerical studies are conducted to compare the proposed DTL and its ensembles with tree-based counterparts. The DTL methods show competitive performances in various settings, and particularly for smooth functionals the DTL demonstrates its superiority over others. Secondly, we propose a collaborative gradient boosting (CGB) algorithm, which can leverage the full potential of boosting multiple types of base learners by using the intrinsic regularization. Inspired by the idea of coopetition (i.e., cooperation and competition), we allow the base learners of different types to collaborate with as well as competing against each other in gradient descent steps. In the CGB algorithm, different types of base learners typically learn from each other toward better model accuracy, i.e., poor-performing base learners would approach better ones. Extensive numerical studies show that the CGB dramatically improves the prediction accuracy by utilizing different types of learners. Our method yields a general strategy that performs better on real data from a wide range of domains. Finally, we introduce a new concept of the average holding price (AHP) in stock market. We show that, under certain assumptions on the investors’ behavior, the average holding price of a stock can be estimated, based on the historical trading prices and volumes. In contrast to the moving average of the stock prices, the AHP can serve as a benchmark for estimating the average profit/loss level of the stock holders. The numerical algorithm for the AHP depends on an recursive equation, which enables us to compute the real-time AHP from any time point of the stock trading. As illustrated in the examples, some trading strategies can also be built upon the AHP.
DegreeDoctor of Philosophy
SubjectMachine learning
Dept/ProgramStatistics and Actuarial Science
Persistent Identifierhttp://hdl.handle.net/10722/298879

 

DC FieldValueLanguage
dc.contributor.advisorYin, G-
dc.contributor.authorLiu, Yehong-
dc.contributor.author刘业鸿-
dc.date.accessioned2021-04-16T11:16:36Z-
dc.date.available2021-04-16T11:16:36Z-
dc.date.issued2021-
dc.identifier.citationLiu, Y. [刘业鸿]. (2021). Supervised machine learning with Delaunay triangulation learner and collaborative gradient boosting. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/298879-
dc.description.abstractIn this thesis, we discuss some statistical machine learning methods and their ensemble learning approachs with theories and applications. Firstly, we propose a new piecewise linear learner, called the Delaunay triangulation learner (DTL). Based on the data samples in a p-dimensional feature space, the Delaunay triangulation algorithm provides a unique triangulation of the space. The triangulation separates the convex hull of the samples into a series of disjoint p-simplices, where the samples are the vertices of the p-simplices. The DTL is constructed by fitting the responses through linear interpolation functions on each of the Delaunay simplices, and thus it approximates the whole functional by a piecewise linear function. In its ensemble learning approaches, we propose the bagging DTLs, random crystal and the boosting DTL, where the DTLs are constructed on the subspaces of the features, and the feature interactions are captured by Delaunay triangle meshes. Extensive numerical studies are conducted to compare the proposed DTL and its ensembles with tree-based counterparts. The DTL methods show competitive performances in various settings, and particularly for smooth functionals the DTL demonstrates its superiority over others. Secondly, we propose a collaborative gradient boosting (CGB) algorithm, which can leverage the full potential of boosting multiple types of base learners by using the intrinsic regularization. Inspired by the idea of coopetition (i.e., cooperation and competition), we allow the base learners of different types to collaborate with as well as competing against each other in gradient descent steps. In the CGB algorithm, different types of base learners typically learn from each other toward better model accuracy, i.e., poor-performing base learners would approach better ones. Extensive numerical studies show that the CGB dramatically improves the prediction accuracy by utilizing different types of learners. Our method yields a general strategy that performs better on real data from a wide range of domains. Finally, we introduce a new concept of the average holding price (AHP) in stock market. We show that, under certain assumptions on the investors’ behavior, the average holding price of a stock can be estimated, based on the historical trading prices and volumes. In contrast to the moving average of the stock prices, the AHP can serve as a benchmark for estimating the average profit/loss level of the stock holders. The numerical algorithm for the AHP depends on an recursive equation, which enables us to compute the real-time AHP from any time point of the stock trading. As illustrated in the examples, some trading strategies can also be built upon the AHP.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshMachine learning-
dc.titleSupervised machine learning with Delaunay triangulation learner and collaborative gradient boosting-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineStatistics and Actuarial Science-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2021-
dc.identifier.mmsid991044360596703414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats