# The long missing python library for confidence intervals
![logo](logo.png)
[![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)
![Build Status](https://github.com/jacobgil/confidenceinterval/workflows/Tests/badge.svg)
[![Downloads](https://static.pepy.tech/personalized-badge/confidenceinterval?period=month&units=international_system&left_color=black&right_color=brightgreen&left_text=Monthly%20Downloads)](https://pepy.tech/project/confidenceinterval)
[![Downloads](https://static.pepy.tech/personalized-badge/confidenceinterval?period=total&units=international_system&left_color=black&right_color=blue&left_text=Total%20Downloads)](https://pepy.tech/project/confidenceinterval)
`pip install confidenceinterval`
This is a package that computes common machine learning metrics like F1, and returns their confidence intervals.
⭐ Very easy to use, with the standard scikit-learn naming convention and interface.
⭐ Support for many metrics, with modern confidence interval methods.
⭐ The only package with analytical computation of the CI for Macro/Micro/Binary averaging F1, Precision and Recall.
⭐ Support for both analytical computation of the confidence intervals, and bootstrapping methods.
⭐ Easy to use interface to compute confidence intervals on new metrics that don't appear here, with bootstrapping.
## The motivation
A confidence interval gives you a lower and upper bound on your metric. It's affected by the sample size, and by how sensitive the metric is to changes in the data.
When making decisions based on metrics, you should prefer narrow intervals. If the interval is wide, you can't be confident that your high performing metric is not just by luck.
While confidence intervals are commonly used by statisticans, with many great R language implementations,
they are astonishingly rarely used by python users, although python took over the data science world !
Part of this is because there were no simple to use python packages for this.
## Getting started
```python
# All the possible imports:
from confidenceinterval import roc_auc_score
from confidence interval import precision_score, recall_score, f1_score
from confidence interval import accuracy_score,
ppv_score,
npv_score,
tpr_score,
fpr_score,
tnr_score
from confidenceinterval.bootstrap import bootstrap_ci
# Analytic CI:
auc, ci = roc_auc_score(y_true,
y_pred,
confidence_level=0.95)
# Bootstrap CI:
auc, ci = roc_auc_score(y_true,
y_pred,
confidence_level=0.95,
method='bootstrap_bca',
n_resamples=5000)
```
## All methods do an analytical computation by default, but can do bootsrapping instead
By default all the methods return an analytical computation of the confidence interval (CI).
For a bootstrap computation of the CI for any of the methods belonw, just specify method='bootstrap_bca', or method='bootstrap_percentile' or method='bootstrap_basic'.
These are different ways of doing the bootstrapping, but method='bootstrap_bca' is the generalibly reccomended method.
You can also pass the number of bootstrap resamples (n_resamples), and a random generator for controling the reproducability:
```python
random_state = np.random.default_rng()
n_resamples=9999
```
## Support for binary, macro and micro averaging for F1, Precision and Recall.
```python
from confidence interval import precision_score, recall_score, f1_score
binary_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='binary')
macro_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='macro')
micro_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='micro')
bootstrap_binary_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='binary', method='bootstrap_bca', n_resamples=5000)
```
The analytical computation here is using the (amazing) 2022 paper of Takahashi et al (reference below).
The paper derived recall and precision only for micro averaging.
We derive the recall and precision confidence intervals for macro F1 as well using the delta method.
## ROC AUC
```python
from confidence interval import roc_auc_score
```
The analytical computation here is a fast implementation of the DeLong method.
## Binary metrics
```python
from confidence interval import accuracy_score,
ppv_score,
npv_score,
tpr_score,
fpr_score,
tnr_score
# Wilson is used by default:
ppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='wilson')
ppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='jeffreys')
ppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='agresti_coull')
ppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='bootstrap_bca')
```
For these methods, the confidence interval is estimated by treating the ratio as a binomial proportion,
see the [wiki page](https://en.wikipedia.org/wiki/Binomial_proportion_confidence_interval).
By default method='wilson', the wilson interval, which behaves better for smaller samples.
method can be one of ['wilson', 'normal', 'agresti_coull', 'beta', 'jeffreys', 'binom_test'], or one of the boostrap methods.
## Get a confidence interval for any custom metric with Bootstrapping
With the bootstrap_ci method, you can get the CI for any metric function that gets y_true and y_pred as arguments.
As an example, lets get the CI for the balanced accuracy metric from scikit-learn.
```python
from confidenceinterval.bootstrap import bootstrap_ci
# You can specify a random generator for reproducability, or pass None
random_generator = np.random.default_rng()
bootstrap_ci(y_true=y_true,
y_pred=y_pred,
metric=sklearn.metrics.balanced_accuracy_score,
confidence_level=0.95,
n_resamples=9999,
method='bootstrap_bca',
random_state=random_generator)
```
----------
Citation
If you use this for research, please cite. Here is an example BibTeX entry:
```
@misc{jacobgildenblatconfidenceinterval,
title={A python library for confidence intervals},
author={Jacob Gildenblat},
year={2023},
publisher={GitHub},
howpublished={\url{https://github.com/jacobgil/confidenceinterval}},
}
```
----------
## References
The binomial confidence interval computation uses the statsmodels package:
https://www.statsmodels.org/dev/generated/statsmodels.stats.proportion.proportion_confint.html
Yandex data school implementation of the fast delong method:
https://github.com/yandexdataschool/roc_comparison
https://ieeexplore.ieee.org/document/6851192
X. Sun and W. Xu, "Fast Implementation of DeLong’s Algorithm for Comparing the Areas Under Correlated Receiver Operating Characteristic Curves," in IEEE Signal Processing Letters, vol. 21, no. 11, pp. 1389-1393, Nov. 2014, doi: 10.1109/LSP.2014.2337313.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8936911/#APP2
Confidence interval for micro-averaged F1 and macro-averaged F1 scores
`Kanae Takahashi,1,2 Kouji Yamamoto,3 Aya Kuchiba,4,5 and Tatsuki Koyama6`
B. Efron and R. J. Tibshirani, An Introduction to the Bootstrap, Chapman & Hall/CRC, Boca Raton, FL, USA (1993)
http://users.stat.umn.edu/~helwig/notes/bootci-Notes.pdf
`Nathaniel E. Helwig, “Bootstrap Confidence Intervals”`
Bootstrapping (statistics), Wikipedia, https://en.wikipedia.org/wiki/Bootstrapping_%28statistics%29
Raw data
{
"_id": null,
"home_page": "https://github.com/jacobgil/confidenceinterval",
"name": "confidenceinterval",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.6",
"maintainer_email": "",
"keywords": "",
"author": "Jacob Gildenblat",
"author_email": "jacob.gildenblat@gmail.com",
"download_url": "https://files.pythonhosted.org/packages/be/27/ce5416af16f455ec977e87aa8951329cd807123bfbe1f5a00f16d52fe5aa/confidenceinterval-1.0.4.tar.gz",
"platform": null,
"description": "# The long missing python library for confidence intervals\n![logo](logo.png)\n\n[![License: MIT](https://img.shields.io/badge/License-MIT-yellow.svg)](https://opensource.org/licenses/MIT)\n![Build Status](https://github.com/jacobgil/confidenceinterval/workflows/Tests/badge.svg)\n[![Downloads](https://static.pepy.tech/personalized-badge/confidenceinterval?period=month&units=international_system&left_color=black&right_color=brightgreen&left_text=Monthly%20Downloads)](https://pepy.tech/project/confidenceinterval)\n[![Downloads](https://static.pepy.tech/personalized-badge/confidenceinterval?period=total&units=international_system&left_color=black&right_color=blue&left_text=Total%20Downloads)](https://pepy.tech/project/confidenceinterval)\n\n`pip install confidenceinterval`\n\nThis is a package that computes common machine learning metrics like F1, and returns their confidence intervals.\n\n\n\u2b50 Very easy to use, with the standard scikit-learn naming convention and interface.\n\n\u2b50 Support for many metrics, with modern confidence interval methods.\n\n\u2b50 The only package with analytical computation of the CI for Macro/Micro/Binary averaging F1, Precision and Recall.\n\n\u2b50 Support for both analytical computation of the confidence intervals, and bootstrapping methods.\n\n\u2b50 Easy to use interface to compute confidence intervals on new metrics that don't appear here, with bootstrapping.\n\n## The motivation\n\nA confidence interval gives you a lower and upper bound on your metric. It's affected by the sample size, and by how sensitive the metric is to changes in the data.\n\nWhen making decisions based on metrics, you should prefer narrow intervals. If the interval is wide, you can't be confident that your high performing metric is not just by luck.\n\nWhile confidence intervals are commonly used by statisticans, with many great R language implementations,\nthey are astonishingly rarely used by python users, although python took over the data science world !\n\nPart of this is because there were no simple to use python packages for this.\n\n\n## Getting started\n\n```python\n# All the possible imports:\nfrom confidenceinterval import roc_auc_score\nfrom confidence interval import precision_score, recall_score, f1_score\nfrom confidence interval import accuracy_score,\n ppv_score,\n npv_score,\n tpr_score,\n fpr_score,\n tnr_score\nfrom confidenceinterval.bootstrap import bootstrap_ci\n\n\n# Analytic CI:\nauc, ci = roc_auc_score(y_true,\n y_pred,\n confidence_level=0.95)\n# Bootstrap CI:\nauc, ci = roc_auc_score(y_true,\n y_pred,\n confidence_level=0.95,\n method='bootstrap_bca',\n n_resamples=5000)\n\n\n\n```\n\n## All methods do an analytical computation by default, but can do bootsrapping instead\nBy default all the methods return an analytical computation of the confidence interval (CI).\n\nFor a bootstrap computation of the CI for any of the methods belonw, just specify method='bootstrap_bca', or method='bootstrap_percentile' or method='bootstrap_basic'.\nThese are different ways of doing the bootstrapping, but method='bootstrap_bca' is the generalibly reccomended method.\n\nYou can also pass the number of bootstrap resamples (n_resamples), and a random generator for controling the reproducability:\n\n```python\nrandom_state = np.random.default_rng()\nn_resamples=9999\n```\n\n## Support for binary, macro and micro averaging for F1, Precision and Recall.\n```python\nfrom confidence interval import precision_score, recall_score, f1_score\nbinary_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='binary')\nmacro_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='macro')\nmicro_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='micro')\nbootstrap_binary_f1, ci = f1_score(y_true, y_pred, confidence_interval=0.95, average='binary', method='bootstrap_bca', n_resamples=5000)\n\n```\n\nThe analytical computation here is using the (amazing) 2022 paper of Takahashi et al (reference below).\nThe paper derived recall and precision only for micro averaging.\nWe derive the recall and precision confidence intervals for macro F1 as well using the delta method.\n\n\n## ROC AUC\n```python\nfrom confidence interval import roc_auc_score\n```\nThe analytical computation here is a fast implementation of the DeLong method.\n\n\n## Binary metrics\n```python\nfrom confidence interval import accuracy_score,\n ppv_score,\n npv_score,\n tpr_score,\n fpr_score,\n tnr_score\n# Wilson is used by default:\nppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='wilson')\nppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='jeffreys')\nppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='agresti_coull')\nppv, ci = ppv_score(y_true, y_pred, confidence_level=0.95, method='bootstrap_bca')\n\n```\n\nFor these methods, the confidence interval is estimated by treating the ratio as a binomial proportion,\nsee the [wiki page](https://en.wikipedia.org/wiki/Binomial_proportion_confidence_interval).\n\nBy default method='wilson', the wilson interval, which behaves better for smaller samples.\n\nmethod can be one of ['wilson', 'normal', 'agresti_coull', 'beta', 'jeffreys', 'binom_test'], or one of the boostrap methods.\n\n## Get a confidence interval for any custom metric with Bootstrapping\nWith the bootstrap_ci method, you can get the CI for any metric function that gets y_true and y_pred as arguments.\n\nAs an example, lets get the CI for the balanced accuracy metric from scikit-learn.\n\n```python\nfrom confidenceinterval.bootstrap import bootstrap_ci\n# You can specify a random generator for reproducability, or pass None\nrandom_generator = np.random.default_rng()\nbootstrap_ci(y_true=y_true,\n y_pred=y_pred,\n metric=sklearn.metrics.balanced_accuracy_score,\n confidence_level=0.95,\n n_resamples=9999,\n method='bootstrap_bca',\n random_state=random_generator)\n```\n\n\n\n----------\n\nCitation\nIf you use this for research, please cite. Here is an example BibTeX entry:\n\n```\n@misc{jacobgildenblatconfidenceinterval,\n title={A python library for confidence intervals},\n author={Jacob Gildenblat},\n year={2023},\n publisher={GitHub},\n howpublished={\\url{https://github.com/jacobgil/confidenceinterval}},\n}\n```\n\n----------\n\n## References\n\nThe binomial confidence interval computation uses the statsmodels package:\nhttps://www.statsmodels.org/dev/generated/statsmodels.stats.proportion.proportion_confint.html\n\nYandex data school implementation of the fast delong method:\nhttps://github.com/yandexdataschool/roc_comparison\n\nhttps://ieeexplore.ieee.org/document/6851192\nX. Sun and W. Xu, \"Fast Implementation of DeLong\u2019s Algorithm for Comparing the Areas Under Correlated Receiver Operating Characteristic Curves,\" in IEEE Signal Processing Letters, vol. 21, no. 11, pp. 1389-1393, Nov. 2014, doi: 10.1109/LSP.2014.2337313.\n\nhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC8936911/#APP2\n\nConfidence interval for micro-averaged F1 and macro-averaged F1 scores\n`Kanae Takahashi,1,2 Kouji Yamamoto,3 Aya Kuchiba,4,5 and Tatsuki Koyama6`\n\nB. Efron and R. J. Tibshirani, An Introduction to the Bootstrap, Chapman & Hall/CRC, Boca Raton, FL, USA (1993)\n\nhttp://users.stat.umn.edu/~helwig/notes/bootci-Notes.pdf\n`Nathaniel E. Helwig, \u201cBootstrap Confidence Intervals\u201d`\n\n\nBootstrapping (statistics), Wikipedia, https://en.wikipedia.org/wiki/Bootstrapping_%28statistics%29\n",
"bugtrack_url": null,
"license": "",
"summary": "Confidence Intervals in python",
"version": "1.0.4",
"project_urls": {
"Bug Tracker": "https://github.com/jacobgil/confidenceinterval/issues",
"Homepage": "https://github.com/jacobgil/confidenceinterval"
},
"split_keywords": [],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "be27ce5416af16f455ec977e87aa8951329cd807123bfbe1f5a00f16d52fe5aa",
"md5": "29b134a22a648cede0fe9c18055ee899",
"sha256": "159af90d9135f67f6140a40728a7d87ade875d0aec9d52da6b0f565d81a51523"
},
"downloads": -1,
"filename": "confidenceinterval-1.0.4.tar.gz",
"has_sig": false,
"md5_digest": "29b134a22a648cede0fe9c18055ee899",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.6",
"size": 15307,
"upload_time": "2023-06-19T03:09:46",
"upload_time_iso_8601": "2023-06-19T03:09:46.126040Z",
"url": "https://files.pythonhosted.org/packages/be/27/ce5416af16f455ec977e87aa8951329cd807123bfbe1f5a00f16d52fe5aa/confidenceinterval-1.0.4.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-06-19 03:09:46",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "jacobgil",
"github_project": "confidenceinterval",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"requirements": [],
"lcname": "confidenceinterval"
}