• 统计学_Wilcoxon signed-rank test(python脚本)


    机器学习,统计项目联系QQ:231469242

     

    两个配对样本,均匀分布,非正太分布

    Wilcoxon signed-rank test

    曼-惠特尼U检验Mann–Whitney Test

    两个独立样本,均匀分布,非正太分布

    https://en.wikipedia.org/wiki/Frank_Wilcoxon

    ranking method

    这是a=0.05对应得样本数

    critical value 对应样本数,20样本对应的关键值为52

    positive sum:

    差异为正值的排名和

    negative sum:

    差异为负值的排名和

    如果negative sum小于52,表明两组数有显著差异,推翻原假设

     

    http://blog.sina.com.cn/s/blog_4bcf5ebd0101422n.html

    SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

    一、概述

    非参数检验对于总体分布没有要求,因而使用范围更广泛。对于两配对样本的非参数检验,首选Wilcoxon符号秩检验。它与配对样本t检验相对应。

     

    二、问题

    为了研究某放松方法(如听音乐)对于入睡时间的影响,选择了10名志愿者,分别记录未进行放松时的入睡时间及放松后的入睡时间(单位为分钟),数据如下笔。请问该放松方法对入睡时间有无影响。

    本例可以采用配对样本t检验,但由于样本量少,数据可能不符合正太分布,所以考虑用非参数检验。

     

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    三、统计操作

     

    数据视图

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    菜单选择

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    打开如下的对话框

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    该对话框有三个选项卡,第一个选项卡会根据第三个选项卡的设置自动设置,故一般不用手动设定。点击进入“字段”选项卡。将“放松前”、“放松后”均选入右边“检验字段”框中。

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    点击进入“设置”对话框,选择检验方法,切换为“自定义检验”,选择“Wilcoxon匹配样本对符号秩(二样本)”复选框。“检验选项”可以设定显著性水平。

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     

    点击“运行”按钮,输出结果

     

    四、结果解读
    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

    这就是输出结果。原假设示放松前好放松后差值的中位数等于0,P=0.015<0.05,拒绝原假设,认为放松前后有统计学差异。

    双击该表格,会弹出如下的“模型浏览器”窗口,可以看到更详细的信息。如下图。

    [转载]SPSS学习笔记之——两配对样本的非参数检验(Wilcoxon符号秩检验)

     # -*- coding: utf-8 -

    import scipy.stats as stats

     

    data1=[21,12,12,23,19,13,20,17,14,19]

    data2=[12,11,8,9,10,15,16,17,10,16]

     

     

    stats.wilcoxon(data1,data2)

     

    '''

    Out[2]: WilcoxonResult(statistic=2.0, pvalue=0.01471359242280415)

    p值小于0.05,两组数据有显著差异

    '''

     

     

     

    https://study.com/academy/lesson/non-parametric-inferential-statistics-definition-examples.html

    In this lesson, you're going to learn about the major differences between parametric and non-parametric methods for dealing with inferential statistics, as well as see an example of the non-parametric method.

    Normal

    What is normal? At least in the world of statistics, this has nothing to do with how someone dresses, acts, or what their beliefs are. Normal data comes from a population with a normal distribution. A normal distribution is a distribution that has a symmetrical bell-shaped curve to it, which you're probably well aware of.

    Keep this concept in mind as we go over the major differences between parametric and non-parametric statistics.

    Parametric Methods

    First, let's define our terms really simply. When we talk about parameters in statistics, what are we actually hinting at? Parameters are descriptive measures of population data, such as the population mean or population standard deviation.

    When the variable we are considering is approximately (or completely) normally distributed, or the sample size is large, we can use two inferential methods that are concerned with parameters - appropriately called parametric methods - when performing a hypothesis test for a population mean. For instance, if we find that the distribution of the average salary of a sample looks like a bell curve, then parametric methods may be used.

    These two methods are probably ones you've heard of before. They are the z-test, which we'd use when the population standard deviation is known to us; or the t-test, which we'd use when the population standard deviation is not known to us.

    Non-Parametric Methods

    Inferential methods that are not concerned with parameters are known, easily enough, as non-parametric methods. However, this term is also more broadly used to refer to many methods that are applied without assuming normality. So, for instance, if we find that the distribution of the average salary of a sample looks like the histogram you see on the screen now [see video], which is nothing close to that of a bell curve, then we will have to turn to non-parametric methods.

    Such non-parametric methods have their pros and cons. On the pro side, these methods are usually simpler to compute and are more resistant to extreme values when compared to parametric methods. On the con side, if the requirements for the use of a parametric method are actually met, non-parametric methods do not have as much power as the z-test or t-test.

    By power, I simply mean the probability of avoiding a type II error, which is an error where we fail to reject the false null hypothesis.

    Example of a Non-Parametric Method

    One example of a non-parametric method is the Wilcoxon signed-rank test. This is a test that assumes the variable under consideration does not need a specific shape and doesn't have to be normally distributed, but is symmetric in its distribution nonetheless. This means that it can be sliced in half to produce two mirror images.

    So, for example, a right-skewed or left-skewed distribution would not be appropriate for this test since it's not symmetric. But a normal, symmetric bimodal, triangular, or uniform distribution would be a fit for this test since any one of those can be sliced in half to produce two mirror images of one another.

    Other non-parametric tests include the likes of:

    • The Kruskal-Wallis test
    • The Mann-Whitney U test
    • The sign test

    Lesson Summary

    Normal data comes from a population with a normal distribution. A normal distribution is a distribution that has a symmetrical bell-shaped curve to it, which you're probably well aware of.

    Inferential methods that are concerned with parameters are appropriately called parametric methods, and include the z-test and t-test. Parameters are descriptive measures of population data.

    Inferential methods that are not concerned with parameters are known as non-parametric methods. This term is also more broadly used to refer to many methods that are applied without assuming normality.

    While non-parametric methods are easier to compute than parametric ones, they do not have as much power as parametric methods if the requirements for the use of a parametric method are met. By power, I simply mean the probability of avoiding a type II error, which is an error where we fail to reject the false null hypothesis.

    An example of a non-parametric method is the Wilcoxon signed-rank test. This is a test that assumes the variable under consideration does not need a specific shape and doesn't have to be normally distributed, but is symmetric in its distribution nonetheless.

     

     https://study.163.com/provider/400000000398149/index.htm?share=2&shareId=400000000398149( 欢迎关注博主主页,学习python视频资源,还有大量免费python经典文章)

     

     

     
  • 相关阅读:
    win_tc使用感受
    10进制转8进制(栈操作)
    动态栈
    数组
    单链表学习
    static用法
    基础2
    linux c first
    linux net command /uboot command
    opencv
  • 原文地址:https://www.cnblogs.com/webRobot/p/6820657.html
Copyright © 2020-2023  润新知