首页 | 本学科首页   官方微博 | 高级检索  
     检索      

决策树剪枝方法的比较
引用本文:魏红宁.决策树剪枝方法的比较[J].西南交通大学学报,2005,40(1):44-48.
作者姓名:魏红宁
作者单位:西南交通大学校长办公室,四川,成都,610031
摘    要:为在决策树剪枝中正确选择剪枝方法,基于理论分析和算例详细地比较了当前主要的4种剪枝方法的计算复杂性、剪枝方式、误差估计和理论基础.与PEP相比,MEP产生的树精度较小且树较大;REP是最简单的剪枝方法之一,但需要独立剪枝集;在同样精度情况下,CCP比REP产生的树小.如果训练数据集丰富,可以选择REP,如果训练数据集较少且剪枝精度要求较高,则可以选用PEP.

关 键 词:数据挖掘  决策树  事后剪枝  PEP  MEP  REP  CCP
文章编号:0258-2724(2005)01-0044-05

Comparison among Methods of Decision Tree Pruning
WEI Hong-ning.Comparison among Methods of Decision Tree Pruning[J].Journal of Southwest Jiaotong University,2005,40(1):44-48.
Authors:WEI Hong-ning
Abstract:To select a suitable pruning method in decision tree pruning, four well-known pruning methods were compared in terms of computational complexity, traversal strategy, error estimation and theoretical principle by taking a classification and regression tree as an example. Compared with pessimistic error pruning (PEP), minimum error pruning (MEP) is less accurate and produces a larger tree. Reduced error pruning (REP) is one of the simplest pruning strategies, but it has the disadvantage of requiring a separate data set for pruning. Cost-complexity pruning (CCP) produces a smaller tree than REP with similar accuracy. Practically, if the training data is abundant, REP is preferable; and if the train data is the expected accuracy is high but with limited data, PEP is good choice.
Keywords:data mining  decision tree  post pruning  pessimistic error pruning  minimum error pruning  reduced error pruning  cost-complexity pruning
本文献已被 CNKI 维普 万方数据 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号