0


多元绩效指标的支持向量法

A Support Vector Method for Multivariate Performance Measures
课程网址: http://videolectures.net/icml05_joachims_pgp/  
主讲教师: Thorsten Joachims
开课单位: 康奈尔大学
开课时间: 2007-04-12
课程语种: 英语
中文简介:
我们研究了不同学习算法的预测与真实后验概率之间的关系。我们证明了最大边缘法,如增强树和增强树桩,将概率质量从0和1推开,在预测概率中产生特征性的乙状扭曲。像NaiveBayes这样的模型做出了不切实际的独立性假设,将概率推向0和1。其他的模型,如神经网络和袋装树,没有这些偏差,并预测校准好的概率。我们用两种方法来修正某些学习方法预测的偏差概率:平标度和等渗回归。我们定性地检查这些校准方法适用于哪些类型的畸变,并定量地检查它们需要多少数据才能有效。实验结果表明,经过校正后,提高了树木、随机森林和支持向量机预测的最佳概率。
课程简介: We examine the relationship between the predictions made by different learning algorithms and true posterior probabilities. We show that maximum margin methods such as boosted trees and boosted stumps push probability mass away from 0 and 1 yielding a characteristic sigmoid shaped distortion in the predicted probabilities. Models such as Naive Bayes, which make unrealistic independence assumptions, push probabilities toward 0 and 1. Other models such as neural nets and bagged trees do not have these biases and predict well calibrated probabilities. We experiment with two ways of correcting the biased probabilities predicted by some learning methods: Platt Scaling and Isotonic Regression. We qualitatively examine what kinds of distortions these calibration methods are suitable for and quantitatively examine how much data they need to be effective. The empirical results show that after calibration boosted trees, random forests, and SVMs predict the best probabilities.
关 键 词: 预测概率; 朴素贝叶斯; 神经网络
课程来源: 视频讲座网
最后编审: 2021-02-03:nkq
阅读次数: 33