线性子空间
决策树
计算机科学
随机森林
一般化
人工智能
随机子空间法
机器学习
模式识别(心理学)
特征(语言学)
特征向量
构造(python库)
树(集合论)
数据挖掘
支持向量机
数学
数学分析
语言学
哲学
几何学
程序设计语言
标识
DOI:10.1109/icdar.1995.598994
摘要
Decision trees are attractive classifiers due to their high execution speed. But trees derived with traditional methods often cannot be grown to arbitrary complexity for possible loss of generalization accuracy on unseen data. The limitation on complexity usually means suboptimal accuracy on training data. Following the principles of stochastic modeling, we propose a method to construct tree-based classifiers whose capacity can be arbitrarily expanded for increases in accuracy for both training and unseen data. The essence of the method is to build multiple trees in randomly selected subspaces of the feature space. Trees in, different subspaces generalize their classification in complementary ways, and their combined classification can be monotonically improved. The validity of the method is demonstrated through experiments on the recognition of handwritten digits.
科研通智能强力驱动
Strongly Powered by AbleSci AI