当前位置: X-MOL 学术Artif. Intell. Rev. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Selective ensemble of uncertain extreme learning machine for pattern classification with missing features
Artificial Intelligence Review ( IF 12.0 ) Pub Date : 2020-04-19 , DOI: 10.1007/s10462-020-09836-3
Shibo Jing , Yidan Wang , Liming Yang

Ensemble learning is an effective technique to improve performance and stability compared to single classifiers. This work proposes a selective ensemble classification strategy to handle missing data classification, where an uncertain extreme learning machine with probability constraints is used as individual (or base) classifiers. Then, three selective ensemble frameworks are developed to optimize ensemble margin distributions and aggregate individual classifiers. The first two are robust ensemble frameworks with the proposed loss functions. The third is a sparse ensemble classification framework with the zero-norm regularization, to automatically select the required individual classifiers. Moreover, the majority voting method is applied to produce ensemble classifier for missing data classification. We demonstrate some important properties of the proposed loss functions such as robustness, convexity and Fisher consistency. To verify the validity of the proposed methods for missing data, numerical experiments are implemented on benchmark datasets with missing feature values. In experiments, missing features are first imputed by using expectation maximization algorithm. Numerical experiments are simulated in filled datasets. With different probability lower bounds of classification accuracy, experimental results under different proportion of missing values show that the proposed ensemble methods have better or comparable generalization compared to the traditional methods in handling missing-value data classifications.

中文翻译:

用于特征缺失模式分类的不确定极限学习机的选择性集成

与单个分类器相比,集成学习是一种提高性能和稳定性的有效技术。这项工作提出了一种选择性集成分类策略来处理缺失数据分类,其中使用具有概率约束的不确定极限学习机作为个体(或基础)分类器。然后,开发了三个选择性集成框架来优化集成边界分布和聚合单个分类器。前两个是具有建议损失函数的稳健集成框架。第三个是具有零范数正则化的稀疏集成分类框架,以自动选择所需的单个分类器。此外,采用多数投票方法产生集成分类器,用于缺失数据分类。我们展示了所提出的损失函数的一些重要特性,例如鲁棒性、凸性和 Fisher 一致性。为了验证所提出的缺失数据方法的有效性,在具有缺失特征值的基准数据集上进行了数值实验。在实验中,首先使用期望最大化算法来估算缺失的特征。在填充数据集中模拟数值实验。在不同的分类准确率概率下界下,不同缺失值比例下的实验结果表明,与传统方法相比,所提出的集成方法在处理缺失值数据分类方面具有更好或相当的泛化能力。为了验证所提出的缺失数据方法的有效性,在具有缺失特征值的基准数据集上进行了数值实验。在实验中,首先使用期望最大化算法来估算缺失的特征。在填充数据集中模拟数值实验。在不同的分类准确率概率下界下,不同缺失值比例下的实验结果表明,与传统方法相比,所提出的集成方法在处理缺失值数据分类方面具有更好或相当的泛化能力。为了验证所提出的缺失数据方法的有效性,在具有缺失特征值的基准数据集上进行了数值实验。在实验中,首先使用期望最大化算法来估算缺失的特征。在填充数据集中模拟数值实验。在不同的分类准确率概率下界下,不同缺失值比例下的实验结果表明,与传统方法相比,所提出的集成方法在处理缺失值数据分类方面具有更好或相当的泛化能力。
更新日期:2020-04-19
down
wechat
bug