Great post!

August 16, 2018

(This article was first published on Stories by Matt.0 on Medium, and kindly contributed to R-bloggers)

Great post!

I wanted to mention that although many previous studies have used the area under receiver operating characteristic curve (auROC) statistic to benchmark the precision, it misleads evaluators when the test data is (highly) imbalanced see: PLOS One, 10(3):e0118432, 2015 & bioRxiv, 2017 doi: 10.1101/142760

In my field (bioinformatics), ENCODE’s DREAM-Challenge recommends reporting both auROC, which assess false negative predictions and the area under precision-recall curve (auPR), which also assesses false positives (see: ENCODE-DREAM in vivo Transcription Factor Binding Site Prediction Challenge., 2017. Accessed: 2018–01–31)

To leave a comment for the author, please follow the link and comment on their blog: Stories by Matt.0 on Medium. offers daily e-mail updates about R news and tutorials on topics such as: Data science, Big Data, R jobs, visualization (ggplot2, Boxplots, maps, animation), programming (RStudio, Sweave, LaTeX, SQL, Eclipse, git, hadoop, Web Scraping) statistics (regression, PCA, time series, trading) and more...

If you got this far, why not subscribe for updates from the site? Choose your flavor: e-mail, twitter, RSS, or facebook...

Comments are closed.

Search R-bloggers


Never miss an update!
Subscribe to R-bloggers to receive
e-mails with the latest R posts.
(You will not see this message again.)

Click here to close (This popup will not appear again)