{"id":1024,"date":"2022-06-14T15:44:48","date_gmt":"2022-06-14T19:44:48","guid":{"rendered":"https:\/\/blog.uvm.edu\/tbplante\/?p=1024"},"modified":"2024-11-06T09:44:59","modified_gmt":"2024-11-06T14:44:59","slug":"descriptive-labels-of-metrics-assessing-discrimination","status":"publish","type":"post","link":"https:\/\/blog.uvm.edu\/tbplante\/2022\/06\/14\/descriptive-labels-of-metrics-assessing-discrimination\/","title":{"rendered":"Descriptive labels of metrics assessing discrimination"},"content":{"rendered":"\n<h2 class=\"wp-block-heading\">Discrimination and calibration of models predicting risk<\/h2>\n\n\n\n<p>Risk prediction is common in medicine, eg the Framingham Risk Score and the Pooled Cohort Equation\/10-year ASCVD risk prediction models. Machine learning models that also predict risk are growing in popularity. <\/p>\n\n\n\n<p>Discrimination and calibration are discussed <a rel=\"noreferrer noopener\" href=\"https:\/\/pubmed.ncbi.nlm.nih.gov\/29049590\/\" target=\"_blank\">in this excellent JAMA article<\/a>. In brief, discrimination relates to how a model divvies up low and high risk individuals. So, in a population of folks of various risk levels, high discriminating models will score higher risk folks higher than low risk folks. For example, a CVD model should say that a 60 year old with hypertension, hyperlipidemia, and diabetes has a higher risk of CVD than a 20 year old with none of those conditions. Calibration, on the other hand, describes how well a model predicts risk for an individual person. <\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Discrimination descriptive labels<\/h2>\n\n\n\n<p>I have to look this up every time, so I am putting this here. Here&#8217;s a widely-used labeling for discrimination, which in <a href=\"https:\/\/www.jmir.org\/2020\/12\/e24048\/\" target=\"_blank\" rel=\"noreferrer noopener\">this manuscript<\/a>, we called it the &#8220;Hosmer-Lemeshow criteria&#8221;. These authors applied this to area under the ROC (AUROC) curves, but I think it&#8217;s also fair to apply to C-statistics.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>0.5: No discrimination<\/li>\n\n\n\n<li>0.5 to &lt;0.7: Poor discrimination<\/li>\n\n\n\n<li>0.7 to &lt;0.8: Acceptable discrimination<\/li>\n\n\n\n<li>0.8 to &lt;0.9: Excellent discrimination<\/li>\n\n\n\n<li>\u22650.9: Outstanding discrimination<\/li>\n<\/ul>\n\n\n\n<p>The reference is here: <\/p>\n\n\n\n<pre class=\"wp-block-preformatted\">Hosmer JD, Lemeshow S, Sturdivant R. Applied Logistic Regression. Hoboken, NJ, USA: John Wiley &amp; Sons Inc; 2013. <\/pre>\n\n\n\n<p>Here&#8217;s the <a rel=\"noreferrer noopener\" href=\"https:\/\/www.google.com\/books\/edition\/Applied_Logistic_Regression\/64JYAwAAQBAJ\" target=\"_blank\">Google Books listing<\/a> for this, in case you want to grab the metadata for this reference using Zotero or whatnot. You&#8217;ll see the above labels on page 177 &#8212; you can see this with the &#8220;Preview&#8221; view.  <\/p>\n","protected":false},"excerpt":{"rendered":"<p>Discrimination and calibration of models predicting risk Risk prediction is common in medicine, eg the Framingham Risk Score and the Pooled Cohort Equation\/10-year ASCVD risk prediction models. Machine learning models that also predict risk are growing in popularity. Discrimination and calibration are discussed in this excellent JAMA article. In brief, discrimination relates to how a &hellip; <a href=\"https:\/\/blog.uvm.edu\/tbplante\/2022\/06\/14\/descriptive-labels-of-metrics-assessing-discrimination\/\" class=\"more-link\">Continue reading <span class=\"screen-reader-text\">Descriptive labels of metrics assessing discrimination<\/span><\/a><\/p>\n","protected":false},"author":4473,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[502558],"tags":[703372,703369,703368,20138,703376,703374,703375,703373,703370],"class_list":["post-1024","post","type-post","status-publish","format-standard","hentry","category-epidemiology-and-biostatistics","tag-accuracy","tag-auroc","tag-c-statistic","tag-discrimination","tag-machine-learning","tag-prediction","tag-prediction-models","tag-receiver-operating-characteristic","tag-roc"],"_links":{"self":[{"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/posts\/1024","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/users\/4473"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/comments?post=1024"}],"version-history":[{"count":3,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/posts\/1024\/revisions"}],"predecessor-version":[{"id":1925,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/posts\/1024\/revisions\/1925"}],"wp:attachment":[{"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/media?parent=1024"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/categories?post=1024"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.uvm.edu\/tbplante\/wp-json\/wp\/v2\/tags?post=1024"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}