{"id":219,"date":"2017-04-18T03:03:33","date_gmt":"2017-04-18T03:03:33","guid":{"rendered":"http:\/\/sites.warnercnr.colostate.edu\/gwhite\/?page_id=219"},"modified":"2017-04-18T03:12:41","modified_gmt":"2017-04-18T03:12:41","slug":"statistical-theory","status":"publish","type":"page","link":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/statistical-theory\/","title":{"rendered":"Statistical Theory"},"content":{"rendered":"<p><span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\"><a href=\"http:\/\/warnercnr.colostate.edu\/~gwhite\/mark\/markhelp\/index.html\">Contents<\/a> &#8211; <a href=\"http:\/\/warnercnr.colostate.edu\/~gwhite\/mark\/markhelp\/idx.htm\">Index<\/a><\/span><\/p>\n<hr \/>\n<p><span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: medium\"><b>Statistical Theory<\/b><\/span><\/p>\n<p><span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">Much of modern statistical theory rests on Fisher&#8217;s likelihood principle.\u00a0 This theory continues to serve as the backbone of general statistical inference across nearly all areas of science.\u00a0 The past two decades have seen an explosion in advanced statistical methods.\u00a0 These advances include\u00a0 generalized linear models (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">McCullagh and Nelder 1989<\/a>) and new theory for the analysis of binary data (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">Cox and Snell 1989<\/a>).<\/span><br \/>\n<span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">\u00a0 Quasi-likelihood methods for modeling the variance-covariance matrix and allowing empirical estimates of sampling variances (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">Wedderburn 1974<\/a>) have been developed and are particularly important in product-multinomial models.\u00a0 Profile likelihood intervals are now frequently used (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">Cormack 1992<\/a>). Use of general information theoretic methods in the selection of a parsimonious model (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">Akaike 1985<\/a>) has been a major theoretical advance.\u00a0 This theory, including Akaike&#8217;s Information Criterion (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/aic-aicc-qaic-aicc\/\">AIC<\/a>) is a substantial advance to the general likelihood theory.<\/span><br \/>\n<span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">Several important statistical advances have resulted because of the greatly increased computational power of relatively inexpensive computers.\u00a0 This has lead to a number of computer intensive statistical methods.\u00a0 The concept of repeated resampling has lead to the bootstrap (<a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/literature-cited\/\">Efron 1982<\/a>) for empirical estimates of the variance-covariance matrix and establishing confidence intervals.\u00a0 Procedures now exist to compute the exact <i>P<\/i>-value for contingency tables where the data are sparse (i.e., the expected values are small).\u00a0 These procedures are relevant for goodness of fit tests for general multinomial models.<\/span><br \/>\n<span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">This body of theory then finds specific application in the analysis of data from capture-recapture and band recovery surveys and experiments.<\/span><\/p>\n<p><span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">More information is available at <\/span><a href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/analysis-marked-animal-encounter-data\/\"><span style=\"color: #0000ff;font-family: 'Courier New', courier, typewriter, monospace;font-size: small\">https:\/\/sites.warnercnr.colostate.edu\/gwhite\/analysis-marked-animal-encounter-data\/<\/span><span style=\"color: #0000ff;font-family: Arial, helvetica, sans-serif;font-size: small\">.<\/span><\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Contents &#8211; Index Statistical Theory Much of modern statistical theory rests on Fisher&#8217;s likelihood principle.\u00a0 This theory continues to serve as the backbone of general statistical inference across nearly all areas of science.\u00a0 The past two decades have seen an &hellip;<\/p>\n<p class=\"read-more\"> <a class=\"more-link\" href=\"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/statistical-theory\/\"> <span class=\"screen-reader-text\">Statistical Theory<\/span> Read More &raquo;<\/a><\/p>\n","protected":false},"author":117,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-219","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/pages\/219","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/users\/117"}],"replies":[{"embeddable":true,"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/comments?post=219"}],"version-history":[{"count":3,"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/pages\/219\/revisions"}],"predecessor-version":[{"id":231,"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/pages\/219\/revisions\/231"}],"wp:attachment":[{"href":"https:\/\/sites.warnercnr.colostate.edu\/gwhite\/wp-json\/wp\/v2\/media?parent=219"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}