同济大学:《大数据分析与数据挖掘 Big Data Analysis and Mining》课程教学资源(PPT课件讲稿)Evaluation & other classifiers

Big Data Analysis and Mining Evaluation other classifiers Qinpei Zhao赵钦佩 qinpeizhao@tongji.edu.cn 2015 Fall 2021/2/8
2021/2/8 1 Big Data Analysis and Mining Qinpei Zhao 赵钦佩 qinpeizhao@tongji.edu.cn 2015 Fall Evaluation & other classifiers

ROC curve Receiver Operating Characteristics(Roc) graphs have long been used in signal detection theory to depict the tradeoff between hit rates and false alarm rates over noisy channel a Recent years have seen an increase in the use of roc graphs in the machine learning community A useful technique for organizing classifiers and visualizing their performance a Especially useful for domains with skewed class distribution and unequal classification error cost
ROC curve ◼ Receiver Operating Characteristics (ROC) graphs have long been used in signal detection theory to depict the tradeoff between hit rates and false alarm rates over noisy channel ◼ Recent years have seen an increase in the use of ROC graphs in the machine learning community ◼ A useful technique for organizing classifiers and visualizing their performance ◼ Especially useful for domains with skewed class distribution and unequal classification error cost

True condition Total population Condition positive Condition negative ∑ Condition positive Positive predictive value(PPv). Predicted condition False positive Precision False discovery rate(FDR) positive 2 True posit posit Test outcome postive condition 2Test outcome positive Predicted condition False negative False omission rate(FOR) Negative predictive value(NPV) (ype II error) True negative 2False negative E True negat ∑ Test outcome negative 2 Test outcome negative True positive rate(IPR) False positive rate(FPR). Positive likelihood ratio (LR+ Sensitivity, Recall FAlse positive curacy (Acc= ∑ Condition negative Diagnostic odds ratio(DOR) 2 True positive+ ETrue negative otal population False negative rate(FNR), True negative rate(TNR), Negative likelihood ratio (LR- ∑ False negative ∑ True negative TNR ∑ Condition negati 2021/2/8
2021/2/8 3

ROC curve ROC curve is a plot of TPR(sensitivity) against FPR(specificity) which depicts relative trade-offs between benefits(true positives) and costs(false positives) number of true positives Sensitivit number of true positives+ number of false negatives number of true negatives Specificity= number of true negatives number of false positives
ROC curve ROC curve is a plot of TPR(sensitivity) against FPR(specificity) which depicts relative trade-offs between benefits (true positives) and costs (false positives)

Exampl e ROC space TP=63FP=2891 P=76‖FP=1288 rfect FN=37TN=72109 0.8 N=24 TN=88112 100 10020(07 100 100200 A Sensitivity 8 o5 better Specificity 0.3 02 c worse 0 0.2 0.4 0.6 0.8 FPR or(1-specificity)
Example

ROC curve a Discrete classifier produces an(FPR, TPR) pair corresponding to a single point in ROC space a Some classifier, such as a naive bayes or a neural network, naturally yield an instance probability or score, a numeric value that represents the degree to which an instance is a member of a class a Such a ranking or scoring classifier can be used with a threshold to produce a discrete classifier a Plotting the Roc point for each possible threshold value results in a curve
ROC curve ◼ A discrete classifier produces an (FPR, TPR) pair corresponding to a single point in ROC space. ◼ Some classifier, such as a Naïve Bayes or a neural network, naturally yield an instance probability or score, a numeric value that represents the degree to which an instance is a member of a class. ◼ Such a ranking or scoring classifier can be used with a threshold to produce a discrete classifier ◼ Plotting the ROC point for each possible threshold value results in a curve

ROC curve 1.0 四≥2 0 A A 巴≥=00 0.6 0.4 0.4 0.2 0.60.81.0 0.2 False positive rate False Positive rate ROC curves show the tradeoff between sensitivity and specificity The closer the curve follows the upper -left border of the Roc space the more accurate the test The closer the curve comes to the 45-degree diagonal of the roc space the less accurate the test a common method is to calculate the area under the roc curve
ROC curve ◼ ROC curves show the tradeoff between sensitivity and specificity. ◼ The closer the curve follows the upper –left border of the ROC space, the more accurate the test. ◼ The closer the curve comes to the 45-degree diagonal of the ROC space, the less accurate the test. ◼ A common method is to calculate the area under the ROC curve

Evaluating a classifier a How well will the classifier we learned perform a novel data? a We can estimate the performance(e.g, accuracy, sensitivity) of the classifier using a test data set Performance on the training data is not a good indicator of performance on future data Test set: independent instances that have not been used in any way to create the classifier Assumption both training data and test data representative samples of the underlying problem
Evaluating a classifier ◼ How well will the classifier we learned perform a novel data? ◼ We can estimate the performance (e.g., accuracy, sensitivity) of the classifier using a test data set ◼ Performance on the training data is not a good indicator of performance on future data ◼ Test set: independent instances that have not been used in any way to create the classifier ◼ Assumption: both training data and test data representative samples of the underlying problem

Holdout Cross-validation method ■ oldout method Given data is randomly partitioned into two independent sets n Training set(e.g, 2/3)for model construction n Test set (e.g, 1/3)for accuracy estimation o Random sampling: a variation of holdout o Repeat holdout k times, accuracy= avg. of the accuracies obtained a Cross-validation(k-fold where ke= 10 is most popular o Randomly partition the data into k mutually exclusive subsets each approximately equal size At i-th iteration, use D as test set and others as training set Leave-one-out k folds where k=# of tuples, for small sized data
Holdout & Cross-validation method ◼ Holdout method ◆ Given data is randomly partitioned into two independent sets Training set (e.g., 2/3) for model construction Test set (e.g., 1/3) for accuracy estimation ◆ Random sampling: a variation of holdout Repeat holdout k times, accuracy = avg. of the accuracies obtained ◼ Cross-validation (k-fold, where k=10 is most popular) ◆ Randomly partition the data into k mutually exclusive subsets, each approximately equal size ◆ At i-th iteration, use Di as test set and others as training set ◆ Leave-one-out: k folds where k = # of tuples, for small sized data

Bootstrap Bootstrap Works well with small data sets Samples the given training tuples uniformly with replacement n i.e., each time a tuple is selected, it is equally likely to be selected again and re-added to the training set a Several bootstrap methods and a common one is. 632 bootstrap o Adata set with d tuples is sampled d times, with replacement resulting in a training set of d samples. The data tuples that did not make it into the training set end up forming the test set. about 63.2%of the original data end up in the bootstrap and the remaining 36.8%form the test set(since(1-1/d)d=e1=0.368) o Repeat the sampling procedure k times, overall accuracy of the model: ACC( M) Xi=10.632*Acc(Mi)testset +0.368* Acc Trainset
Bootstrap ◼ Bootstrap ◆ Works well with small data sets ◆ Samples the given training tuples uniformly with replacement i.e., each time a tuple is selected, it is equally likely to be selected again and re-added to the training set ◼ Several bootstrap methods and a common one is .632 bootstrap ◆ A data set with d tuples is sampled d times, with replacement, resulting in a training set of d samples. The data tuples that did not make it into the training set end up forming the test set. About 63.2% of the original data end up in the bootstrap, and the remaining 36.8% form the test set (since (1-1/d) d =e-1=0.368) ◆ Repeat the sampling procedure k times, overall accuracy of the model: 𝐴𝑐𝑐(𝑀) = 1 𝑘 σ𝑖=1 𝑘 (0.632 ∗ 𝐴𝑐𝑐(𝑀𝑖 )𝑡𝑒𝑠𝑡𝑠𝑒𝑡 + 0.368 ∗ 𝐴𝑐𝑐 𝑀𝑖 𝑡𝑟𝑎𝑖𝑛𝑠𝑒𝑡 )
按次数下载不扣除下载券;
注册用户24小时内重复下载只扣除一次;
顺序:VIP每日次数-->可用次数-->下载券;
- 面积对象编程(PPT讲稿)Object-Oriented Programming and Classes.ppt
- 《计算机网络概述》教学资源(PPT课件讲稿).ppt
- 《计算机组成原理》课程PPT教学课件(讲稿)第三章 计算机核心部件及其工作原理.ppt
- 《大型机系统管理技术》课程教学资源(PPT课件讲稿)第2章 大型服务器外存管理.ppt
- 《ARM嵌入式软件开发》课程教学资源(PPT课件讲稿)第三章 ARM体系结构及编程模型.ppt
- 北京大学:基于信息利用的烟花算法研究(PPT讲稿)Research on Fireworks Algorithms from the Perspective of Information Utilization.pptx
- 系统编程工具REXX和CLIST.ppt
- 《软件测试 Software Testing》教学资源(PPT讲稿)Part 1 The Big Picture.ppt
- 西南民族大学:软件需求分析与总体设计(PPT讲稿,主讲:殷锋).ppt
- 中国地质大学(武汉):R语言入门教程(PPT讲稿).ppt
- 对外经济贸易大学:《大学计算机基础》课程电子教案(PPT课件)第5章 PowerPoint幻灯片制作(PowerPoint 2010).pptx
- 西安培华学院:《计算机网络工程》课程教学资源(PPT课件讲稿)第1章 网络工程知识(主讲:张伟).ppt
- 《计算机网络安全》课程教学资源(PPT课件讲稿)第四章 数据加密技术.ppt
- 计算机应用专业《计算机网络》教学大纲.doc
- 《计算机网络 COMPUTER NETWORKS》课程教学资源(PPT课件讲稿)Chapter 18 互联网协议 Internet Protocols(IP).ppt
- 淮阴工学院:《数据库原理》课程教学资源(PPT课件讲稿)第4章 结构化查询语言SQL.ppt
- 《C++语言程序设计》课程教学资源(PPT课件)第14讲 运算符重载.ppt
- 东南大学:《C++语言程序设计》课程教学资源(PPT课件讲稿)Chapter 13 Object-Oriented Programming - Polymorphism.ppt
- 山东大学:《微机原理及单片机接口技术》课程教学资源(PPT课件讲稿)第四章 指令系统及汇编语言程序设计(4.1-4.4).ppt
- 多媒体技术及应用(PPT讲稿)多媒体音频技术.ppt
- 香港中文大学:Arm board tutorial Part 1 Using the ARM board And start working with C Tutorial 5 and 6.pptx
- 清华大学出版社:《JAVA程序设计实例教程》课程教材电子教案(PPT课件讲稿,共七章,主编:关忠).ppt
- 香港浸会大学:Community Search over Big Graphs:Models, Algorithms, and Opportunities.ppt
- 《数字图像处理》课程教学资源(PPT课件讲稿)第5章 图像编码与压缩.ppt
- 厦门理工学院:《网页设计》培训课件教学资源(PPT课件).ppt
- 西安电子科技大学:《计算机操作系统》课程PPT教学课件(讲稿)第六章 文件管理.ppt
- 机器翻译研讨会(PPT讲稿)神经机器翻译前沿进展(PPT讲稿).pptx
- 山东大学:《微机原理及单片机接口技术》课程教学资源(PPT课件讲稿)第三章 计算机系统的组成与工作原理 3.1 理解模型机的结构及工作过程 3.2 掌握单片机的结构.ppt
- 清华大学出版社:《计算机导论 Introduction to Computer Science》课程配套教材教学资源(PPT课件讲稿,第3版)第4章 操作系统与网络知识.ppt
- 《数据库系统原理》课程PPT教学课件(SQLServer)第7章 Transact-SQL程序设计.ppt
- 《Chemdraw 软件教程》教学资源(PPT讲稿)第一部分 ChemDraw简介.ppt
- 北京大学:计算智能实验室(PPT讲稿)烟花算法算子分析.pptx
- 上饶师范学院:《数据库系统原理 An Introduction to Database System》课程教学资源(PPT课件讲稿,共九章).ppt
- 清华大学:《数据中心网络 Data Center Networking》课程教学资源(PPT课件讲稿).pptx
- 人工神经网络(ANN)方法简介(PPT课件讲稿).ppt
- 合肥工业大学:使用大数据进行计算建模(PPT讲稿)Computing/Modeling with Big Data(主讲:吴信东).pptx
- 《模式识别》课程教学资源(PPT讲稿)Learning with information of features.ppt
- 烟台大学:《C语言程序设计》课程电子教案(PPT课件讲稿)第五章 数组、字符串、指针(主讲:荆蕾).ppt
- 《数据结构》课程教学资源(PPT课件讲稿)第六章 树与二叉树.ppt
- 南京大学:《计算机图形学》课程教学资源(PPT课件讲稿)第6讲 图形观察与几何变换.pptx