期刊文献+

互信息、冗余与广义似然比研究

RESEARCH ON MUTUAL INFORMATION,REDUNDANCY AND GENERALIZED LIKELIHOOD RATIO
下载PDF
导出
摘要 统计学里有很多描述变量间相关性的方法,大部分都要求随机变量必须服从某一或某些概率分布,要么就是满足一定的假设。互信息是基于熵来测量随机变量间的相关性的,它不需要随机变量满足任何特定分布亦或者是特殊的前提假设。一些研究中,冗余也已经作为一种类似于互信息的方法用以评价变量间的相互关系。对冗余和互信息的概念进行深入研究,并使之用以多维的分类数据。研究发现,在几种独立对数线性模型下,分类数据的互信息和冗余可以表示为广义似然比的函数。广义似然比对样本容量是非常敏感的,但是分类数据的互信息和冗余却并不取决于样本容量而是取决于单元概率。因此互信息和冗余可以用作评价分类数据间的相关关系,既不需要特殊的前提假设又不受样本容量的影响。通过示例验证,针对多维数据,冗余又优于互信息。 There are a lot of methods concerning the description of dependency between random variables,most of which oblige random variables either obey one or a few probability distributions or satisfy certain assumptions.Mutual Information bases on entropy to measure the correlation between random variables.It doesn’t require random variables to satisfy any specific distribution or certain assumptions.In some researches,redundancy has already been used as a similar method of mutual information to estimate the relationship between the variables. The thesis explores the concept of redundancy and mutual information,and applies both to multidimensional categorized data.Researches dis-cover that for several independent log-linear models,the mutual information and redundancy of the categorized data can be expressed as a function of general likelihood ratio.General likelihood ratio is very sensitive to the size of data sample,but the mutual information and redun-dancy of the classification data does not depend on the size of data sample but on unit probability.Therefore mutual information and redundan-cy can evaluate the correlation between categorized data,requiring neither specified assumptions nor sample capacity.Case demonstrations validate that,for multidimensional data,redundancy is superior to mutual information.
作者 孟平 包成刚
出处 《计算机应用与软件》 CSCD 北大核心 2014年第2期325-329,共5页 Computer Applications and Software
关键词 互信息 冗余 广义似然比 Entropy Mutual information Redundancy Generalized likelihood ratio
  • 相关文献

参考文献13

  • 1Shannon C E. A mathematical theory of communication[J].{H}Bell System Technical Journal,1948.379423.
  • 2Joe H. Relative entropy measure of multivariate dependence[J].{H}Journal of the American Statistical Association,1989.157164.
  • 3Joe H. Estimation for entropy and other functional of a multivariate density[J].Ann Inst Star Math,1989.683697.
  • 4Darbellay G A. An estimator for the mutual information based on a cri-terion for conditional independence[J].{H}COMPUTATIONAL STATISTICS & DATA ANALYSIS,1999,(01):117.
  • 5Fraser A,Swinney H. Independent coordinates strange attractors from mutual information[J].{H}Physical Review,1986,(02):11341140.
  • 6Palus M. Identifying and quantifying chaos by using information theo-retic function in time series prediction:forecasting the future and un-derstanding the past[J].SantaFe Inst Stud Sci Complexity,1993.387413.
  • 7Prichard D,Theiler J. Generating surrogate data for time series with several simultaneously measured variables[J].{H}Physical Review Letters,1994.951954.
  • 8Wienholt W,Sendhoff B. How to determine the redundancy of noisy chaotic time series[J].{H}International Journal of Bifurcation and Chaos,1996.101117.
  • 9Brillinger R. Some data analyses using mutual information[J].Braz J Probab Stat,2004.163183.
  • 10Brillinger R,Guha A. Mutual information in the frequency domain[J].J Stat Plan Inference,2007,(03):10761084.

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部