ImageVerifierCode 换一换
格式:DOCX , 页数:12 ,大小:257.93KB ,
资源ID:17374256      下载积分:3 金币
快捷下载
登录下载
邮箱/手机:
温馨提示:
快捷下载时,用户名和密码都是您填写的邮箱或者手机号,方便查询和重复下载(系统自动生成)。 如填写123,账号就是123,密码也是123。
特别说明:
请自助下载,系统不会自动发送文件的哦; 如果您已付费,想二次下载,请登录后访问:我的下载记录
支付方式: 支付宝    微信支付   
验证码:   换一换

加入VIP,免费下载
 

温馨提示:由于个人手机设置不同,如果发现不能下载,请复制以下地址【https://www.bdocx.com/down/17374256.html】到电脑端继续下载(重复下载不扣费)。

已注册用户请登录:
账号:
密码:
验证码:   换一换
  忘记密码?
三方登录: 微信登录   QQ登录  

下载须知

1: 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。
2: 试题试卷类文档,如果标题没有明确说明有答案则都视为没有答案,请知晓。
3: 文件的所有权益归上传用户所有。
4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
5. 本站仅提供交流平台,并不能对任何下载内容负责。
6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

版权提示 | 免责声明

本文(JMP和Minitab的比较二简单回归分析文档格式.docx)为本站会员(b****5)主动上传,冰豆网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知冰豆网(发送邮件至service@bdocx.com或直接QQ联系客服),我们立即给予删除!

JMP和Minitab的比较二简单回归分析文档格式.docx

1、将相同的两列数据“X”和“Y”分别输入到最新版的JMP7和Minitab15中,想得到线性回归方程、含回归直线的散点图、回归检验报告以及回归直线的预测区间。对比项目一:操作的便捷性。JMP的操作路径为:主菜单Analyze Fit Y By X,初始报告弹出菜单中的Fit Line,以及Linear Fit弹出报告中的Confid Curve Fit和Confid Curve Indiv等相关选项,得到的报表如图一所示;Minitab的操作路径为:主菜单Stat Regression Fitted Line Plot,Options中选择Display confidence interval和

2、Display prediction interval,得到的报告和图形经整合后如图二所示。操作实现的时间没有明显的差异,但JMP的操作模式让人意识到操作步骤之间层层递进的关系,逻辑性强,而Minitab的操作则纯粹是靠用户用记忆力连接起来的一组相对独立的机械动作。对比项目二:输出报表的整体效果。JMP将统计分析结果和相关图形天然地整合在一起,用户查阅起来一目了然。而Minitab的统计分析结果显示在Session窗口,而相关图形又显示在另一个独立的Graph窗口中,查阅起来平添了几分麻烦。如果分析的数据、内容、次数一多,这种麻烦就更难忍受了。对比项目三:统计分析的具体内容。无论是回归方程的系

3、数,还是R2、显著性检验P值等等,JMP和Minitab的输出结果都是一致的,这说明两种软件背后所遵循的统计原理其实都是一样的。如果观察得更仔细一些,你会发现JMP中的小数位保留得比Minitab更多,而且可以自定义,显得更精确、更专业一些。图二 Minitab的输出结果对比项目四:统计图形的效果。在回归分析的早期,只需要观察最基本的散点图,JMP和Minitab的图形效果差不多。但是到了回归模型的预测应用阶段,置信区间的显示至关重要,JMP可以通过“区间阴影化”的方式加深用户对预测模型的理解。相比之下,Minitab就相形见拙了。如果要比较边际图的效果,两者的差距就更大了。JMP只需在原有的

4、报表基础上再选择Histogram Borders就能完成,结果如图三所示。它既保留了原先预测区间的特征,又能实现其中散点图与直方图之间的动态链接,Minitab则要重新从主菜单中选择Graph Marginal Plot,重新在一个新的Graph窗口才能完成,结果如图四所示。而且可惜的是,原先预测区间的特征消失了,图形之间动态链接的效果更是从来都无法体现的。图三 JMP的边际图图四 Minitab的边际图对比项目五:统计分析的拓展性。JMP和Minitab都考虑到了这一点,但无论是广度,还是深度来看,两者之间的差异都很明显。先看广度,除了两者都具备的功能外,JMP的回归报表中还整合了非参数拟

5、合、样条拟合、分组拟合、特殊拟合和椭圆密度等丰富实用的内容,令Minitab望尘莫及。即使是双方都涉及的内容,我们也可以挖掘其涉及的深度来观察两者的差别。以多项式回归为例,JMP最高可支持六次项,Minitab则仅为三次项。以保存数据为例,JMP不仅能够保存残差值和预测值,而且能够保存预测公式,Minitab则不具备保存公式的功能。诸如此类,举不胜举。唯一可以让Minitab挽回一些脸面的是它在进行残差分析的时候会比JMP稍快一些。总结以上五项对比内容的结果,所有真正理解回归的人都会得到一个一致的结论:JMP在“简单回归分析”方面远胜于Minitab。这个结论的正确性在我们做一些简单的工作时可

6、能会体会不深,但是随着分析问题的深入,这种感觉会越来越强烈地让人感受到。同样,笔者愿以此文抛砖引玉,希望有更多真正理解统计、需要统计来进行质量管理、六西格玛项目的爱好者来交流切磋,共同提高。k-折交叉验证(K-fold cross-validation)是指将样本集分为k份,其中k-1份作为训练数据集,而另外的1份作为验证数据集。用验证集来验证所得分类器或者回归的错误码率。一般需要循环k次,直到所有k份数据全部被选择一遍为止。Cross ValidationCross validation is a model evaluation method that is better than res

7、iduals. The problem with residual evaluations is that they do not give an indication of how well the learner will do when it is asked to make new predictions for data it has not already seen. One way to overcome this problem is to not use the entire data set when training a learner. Some of the data

8、 is removed before training begins. Then when training is done, the data that was removed can be used to test the performance of the learned model on new data. This is the basic idea for a whole class of model evaluation methods called cross validation. The holdout method is the simplest kind of cro

9、ss validation. The data set is separated into two sets, called the training set and the testing set. The function approximator fits a function using the training set only. Then the function approximator is asked to predict the output values for the data in the testing set (it has never seen these ou

10、tput values before). The errors it makes are accumulated as before to give the mean absolute test set error, which is used to evaluate the model. The advantage of this method is that it is usually preferable to the residual method and takes no longer to compute. However, its evaluation can have a hi

11、gh variance. The evaluation may depend heavily on which data points end up in the training set and which end up in the test set, and thus the evaluation may be significantly different depending on how the division is made. K-fold cross validation is one way to improve over the holdout method. The da

12、ta set is divided into k subsets, and the holdout method is repeated k times. Each time, one of the k subsets is used as the test set and the other k-1 subsets are put together to form a training set. Then the average error across all k trials is computed. The advantage of this method is that it mat

13、ters less how the data gets divided. Every data point gets to be in a test set exactly once, and gets to be in a training set k-1 times. The variance of the resulting estimate is reduced as k is increased. The disadvantage of this method is that the training algorithm has to be rerun from scratch k

14、times, which means it takes k times as much computation to make an evaluation. A variant of this method is to randomly divide the data into a test and training set k different times. The advantage of doing this is that you can independently choose how large each test set is and how many trials you a

15、verage over. Leave-one-out cross validation is K-fold cross validation taken to its logical extreme, with K equal to N, the number of data points in the set. That means that N separate times, the function approximator is trained on all the data except for one point and a prediction is made for that

16、point. As before the average error is computed and used to evaluate the model. The evaluation given by leave-one-out cross validation error (LOO-XVE) is good, but at first pass it seems very expensive to compute. Fortunately, locally weighted learners can make LOO predictions just as easily as they

17、make regular predictions. That means computing the LOO-XVE takes no more time than computing the residual error and it is a much better way to evaluate models. We will see shortly that Vizier relies heavily on LOO-XVE to choose its metacodes. Figure 26: Cross validation checks how well a model gener

18、alizes to new dataFig. 26 shows an example of cross validation performing better than residual error. The data set in the top two graphs is a simple underlying function with significant noise. Cross validation tells us that broad smoothing is best. The data set in the bottom two graphs is a complex

19、underlying function with no noise. Cross validation tells us that very little smoothing is best for this data set. Now we return to the question of choosing a good metacode for data set a1.mbl:File - Open - a1.mblEdit - Metacode - A90:9Model - LOOPredict L90: L10:LOOPredict goes through the entire d

20、ata set and makes LOO predictions for each point. At the bottom of the page it shows the summary statistics including Mean LOO error, RMS LOO error, and information about the data point with the largest error. The mean absolute LOO-XVEs for the three metacodes given above (the same three used to gen

21、erate the graphs in fig. 25), are 2.98, 1.23, and 1.80. Those values show that global linear regression is the best metacode of those three, which agrees with our intuitive feeling from looking at the plots in fig. 25. If you repeat the above operation on data set b1.mbl youll get the values 4.83, 4

22、.45, and 0.39, which also agrees with our observations. What are cross-validation and bootstrapping?-Cross-validation and bootstrapping are both methods for estimating generalization error based on resampling (Weiss and Kulikowski 1991; Efron and Tibshirani 1993; Hjorth 1994; Plutowski, Sakata, and

23、White 1994; Shao and Tu 1995). The resulting estimates of generalization error are often used for choosing among various models, such as different network architectures. Cross-validation+In k-fold cross-validation, you divide the data into k subsets of (approximately) equal size. You train the net k

24、 times, each time leaving out one of the subsets from training, but using only the omitted subset to compute whatever error criterion interests you. If k equals the sample size, this is called leave-one-out cross-validation. Leave-v-out is a more elaborate and expensive version of cross-validation t

25、hat involves leaving out all possible subsets of v cases. Note that cross-validation is quite different from the split-sample or hold-out method that is commonly used for early stopping in NNs. In the split-sample method, only a single subset (the validation set) is used to estimate the generalizati

26、on error, instead of k different subsets; i.e., there is no crossing. While various people have suggested that cross-validation be applied to early stopping, the proper way of doing so is not obvious. The distinction between cross-validation and split-sample validation is extremely important because

27、 cross-validation is markedly superior for small data sets; this fact is demonstrated dramatically by Goutte (1997) in a reply to Zhu and Rohwer (1996). For an insightful discussion of the limitations of cross-validatory choice among several learning methods, see Stone (1977). Jackknifing+Leave-one-

28、out cross-validation is also easily confused with jackknifing. Both involve omitting each training case in turn and retraining the network on the remaining subset. But cross-validation is used to estimate generalization error, while the jackknife is used to estimate the bias of a statistic. In the j

29、ackknife, you compute some statistic of interest in each subset of the data. The average of these subset statistics is compared with the corresponding statistic computed from the entire sample in order to estimate the bias of the latter. You can also get a jackknife estimate of the standard error of

30、 a statistic. Jackknifing can be used to estimate the bias of the training error and hence to estimate the generalization error, but this process is more complicated than leave-one-out cross-validation (Efron, 1982; Ripley, 1996, p. 73). Choice of cross-validation method+Cross-validation can be used

31、 simply to estimate the generalization error of a given model, or it can be used for model selection by choosing one of several models that has the smallest estimated generalization error. For example, you might use cross-validation to choose the number of hidden units, or you could use cross-validation to choose a subset of the inputs (subset selection). A subset that conta

copyright@ 2008-2022 冰豆网网站版权所有

经营许可证编号:鄂ICP备2022015515号-1