# Auto Annotator


[](https://codecov.io/gh/CatsWhoTrain/autoannotator)
[](https://badge.fury.io/py/autoannotator)









An extendable tool for automatic annotation of image data by a combination of deep neural networks.
The primary objective of this annotator is to prioritize the accuracy and quality of predictions over speed. The `autoannotator` has been specifically designed to surpass the precision offered by most publicly available tools. It leverages ensembles of deep neural models to ensure the utmost quality in its predictions. It is important to note that neural networks trained on clean datasets tend to yield superior results compared to those trained on larger but noisier datasets.
## Supported tasks
### Human Face Detection and Recognition
### Human Face Detection and Recognition
- [X] Face and landmarks detection
- [X] Face alignment via keypoints
- [X] Face descriptor extraction
### Human Body Detection
- [X] UniHCP
- [X] IterDETR (Progressive DETR)
- [X] RTDETR, ResNet-101
- [X] InternImage-XL
### Other
- [X] DBSCAN clustering
## 📊 Benchmarks
### Face Recognition
#### Speed
| Task | Hardware | Time, s |
| --- | --- | --- |
| Face detection + landmarks + extraction |Xeon e5 2678 v3| ~1 |
### Human Detection
#### Quality
#### AP@50 when RT-DETR ResNet-18 was trained on CUHK-SYSU
| Test set | Original Markup | Markup via DDQ-DETR | Markup via Ensemle |
| ----------- | --------------- | ------------------- | ------------------ |
| CrowdHuman | 52.30 | 76.97 | 77.31 |
| WiderPerson | 54.66 | 60.89 | 63.71 |
## 🏗 Installation
### PIP package
```bash
pip install autoannotator
```
## 🚲 Getting started
### Face recognition example
Check out our demo face recognition pipeline at: `examples/face_recognition_example.py`
#### [Optional] Run frontend and backend
```bash
git clone https://github.com/CatsWhoTrain/autoannotator_client
cd autoannotator_client
docker compose up
```
The webinterface could be found locally at: `http://localhost:8080/`
### Human full-body detection
Detailed description is given in the separate [document](docs/human_detection.md)
#### Human detection example
Check out our demo face recognition pipeline at: `examples/human_detection_example.py`
## FAQ
### Do companies and engineers actually need this tool?
We have asked engineers in the field of video analytics whether they are interested in such a library. Their responses were:
- [IREX](https://irex.ai/): would use this library and contribute to it.
- [NapoleonIT](https://napoleonit.ru/): would use this library and contribute to it.
- [Linza Metrics](https://linzametrics.com/): would use this library.
### What are the reasons for choosing this data labeling tool over the alternative of employing human annotators?
#### Human accuracy is not so good
Long time ago Andrej Karpathy [observed](http://karpathy.github.io/2011/04/27/manually-classifying-cifar10/) that his accuracy was only 94% when he tried to label just 400 images of the CIFAR-10 dataset while SOTA [Efficient adaptive ensembling for image classification](https://onlinelibrary.wiley.com/doi/10.1111/exsy.13424) (August 29, 2023) achieves >99.6% accuracy.
When expert labelers had to choose from ~100 labels while annotating ImageNet, [their error rate increased to 13-15%](http://karpathy.github.io/2014/09/02/what-i-learned-from-competing-against-a-convnet-on-imagenet/).
Andrej's error rate was determined to be 5.1%, and he initially invested approximately one minute in labeling a single image. Conversely, utilizing [Florence](https://arxiv.org/abs/2111.11432v1) or never models for the same task can deliver a top-5 error rate of less than 1%.
##### Industry case: human face classification.
A certain undisclosed company, bound by a non-disclosure agreement (NDA), has utilized a technique wherein face images captured under challenging environmental conditions are pre-processed. This procedure involves the application of both a facial recognition network and DBSCAN algorithm to divide the images into distinct individuals. Subsequently, human annotators undertook a validation process to verify the accuracy of the pre-processed data. The work conducted by the annotators was inspected by their team leader. Ultimately, it was determined by an ML engineer that 1.4% of the clustered face images were mislabeled.
## 🏰 Legacy
Current repository takes ideas and some code form the following projects:
- [faces_detanator](https://github.com/IgorHoholko/faces_detanator)
## ✒ Legal
All individuals depicted in the "assets" images have provided explicit consent to utilize their photos for the purpose of showcasing the current library's work. Kindly refrain from utilizing these images within your individual projects.
Raw data
{
"_id": null,
"home_page": "",
"name": "autoannotator",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.10",
"maintainer_email": "",
"keywords": "",
"author": "",
"author_email": "Alexander FIlonenko <alexander@filonenko.net>, Igor Popov <ig.popov1997@gmail.com>, Roman Savoskin <xromndev@gmail.com>",
"download_url": "https://files.pythonhosted.org/packages/b4/b6/c2197bf171c943b87b23d7b9b6185e5c5c98a77056bbb4a83fddf5e88fd9/autoannotator-0.2.0.tar.gz",
"platform": null,
"description": "# Auto Annotator\n\n\n\n[](https://codecov.io/gh/CatsWhoTrain/autoannotator)\n[](https://badge.fury.io/py/autoannotator)\n\n\n\n\n\n\n\n\n\n\nAn extendable tool for automatic annotation of image data by a combination of deep neural networks.\n\nThe primary objective of this annotator is to prioritize the accuracy and quality of predictions over speed. The `autoannotator` has been specifically designed to surpass the precision offered by most publicly available tools. It leverages ensembles of deep neural models to ensure the utmost quality in its predictions. It is important to note that neural networks trained on clean datasets tend to yield superior results compared to those trained on larger but noisier datasets.\n\n## Supported tasks\n### Human Face Detection and Recognition\n### Human Face Detection and Recognition\n- [X] Face and landmarks detection \n- [X] Face alignment via keypoints\n- [X] Face descriptor extraction\n\n### Human Body Detection\n- [X] UniHCP\n- [X] IterDETR (Progressive DETR)\n- [X] RTDETR, ResNet-101\n- [X] InternImage-XL\n\n### Other\n- [X] DBSCAN clustering\n\n## \ud83d\udcca Benchmarks\n### Face Recognition\n#### Speed\n| Task | Hardware | Time, s |\n| --- | --- | --- |\n| Face detection + landmarks + extraction |Xeon e5 2678 v3| ~1 |\n\n### Human Detection\n#### Quality\n#### AP@50 when RT-DETR ResNet-18 was trained on CUHK-SYSU\n\n| Test set | Original Markup | Markup via DDQ-DETR | Markup via Ensemle |\n| ----------- | --------------- | ------------------- | ------------------ |\n| CrowdHuman | 52.30 | 76.97 | 77.31 |\n| WiderPerson | 54.66 | 60.89 | 63.71 |\n\n\n## \ud83c\udfd7 Installation\n### PIP package\n```bash\npip install autoannotator\n```\n\n## \ud83d\udeb2 Getting started\n### Face recognition example\nCheck out our demo face recognition pipeline at: `examples/face_recognition_example.py` \n\n#### [Optional] Run frontend and backend\n```bash\ngit clone https://github.com/CatsWhoTrain/autoannotator_client\ncd autoannotator_client\ndocker compose up\n```\nThe webinterface could be found locally at: `http://localhost:8080/`\n\n### Human full-body detection\nDetailed description is given in the separate [document](docs/human_detection.md)\n#### Human detection example\nCheck out our demo face recognition pipeline at: `examples/human_detection_example.py` \n\n## FAQ\n### Do companies and engineers actually need this tool?\nWe have asked engineers in the field of video analytics whether they are interested in such a library. Their responses were:\n- [IREX](https://irex.ai/): would use this library and contribute to it.\n- [NapoleonIT](https://napoleonit.ru/): would use this library and contribute to it.\n- [Linza Metrics](https://linzametrics.com/): would use this library.\n\n### What are the reasons for choosing this data labeling tool over the alternative of employing human annotators?\n#### Human accuracy is not so good\nLong time ago Andrej Karpathy [observed](http://karpathy.github.io/2011/04/27/manually-classifying-cifar10/) that his accuracy was only 94% when he tried to label just 400 images of the CIFAR-10 dataset while SOTA [Efficient adaptive ensembling for image classification](https://onlinelibrary.wiley.com/doi/10.1111/exsy.13424) (August 29, 2023) achieves >99.6% accuracy.\nWhen expert labelers had to choose from ~100 labels while annotating ImageNet, [their error rate increased to 13-15%](http://karpathy.github.io/2014/09/02/what-i-learned-from-competing-against-a-convnet-on-imagenet/).\n\nAndrej's error rate was determined to be 5.1%, and he initially invested approximately one minute in labeling a single image. Conversely, utilizing [Florence](https://arxiv.org/abs/2111.11432v1) or never models for the same task can deliver a top-5 error rate of less than 1%.\n\n##### Industry case: human face classification.\nA certain undisclosed company, bound by a non-disclosure agreement (NDA), has utilized a technique wherein face images captured under challenging environmental conditions are pre-processed. This procedure involves the application of both a facial recognition network and DBSCAN algorithm to divide the images into distinct individuals. Subsequently, human annotators undertook a validation process to verify the accuracy of the pre-processed data. The work conducted by the annotators was inspected by their team leader. Ultimately, it was determined by an ML engineer that 1.4% of the clustered face images were mislabeled.\n\n\n\n## \ud83c\udff0 Legacy\nCurrent repository takes ideas and some code form the following projects:\n- [faces_detanator](https://github.com/IgorHoholko/faces_detanator)\n\n## \u2712 Legal\nAll individuals depicted in the \"assets\" images have provided explicit consent to utilize their photos for the purpose of showcasing the current library's work. Kindly refrain from utilizing these images within your individual projects.",
"bugtrack_url": null,
"license": "",
"summary": "Automatic annotation of images library",
"version": "0.2.0",
"project_urls": {
"Bug Tracker": "https://github.com/CatsWhoTrain/autoannotator/issues",
"Homepage": "https://github.com/CatsWhoTrain/autoannotator"
},
"split_keywords": [],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "43a546ee0f07167bcfdde7f36245025903cc0d8b00ed197f8929f73439472101",
"md5": "c21e9957569482c25ff81daf73295e82",
"sha256": "6a98f03077082e6e15b10fcad84a14a9b864f0c97737b29c68bb7164c28ca9f5"
},
"downloads": -1,
"filename": "autoannotator-0.2.0-py3-none-any.whl",
"has_sig": false,
"md5_digest": "c21e9957569482c25ff81daf73295e82",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.10",
"size": 40047,
"upload_time": "2024-01-20T09:21:59",
"upload_time_iso_8601": "2024-01-20T09:21:59.788718Z",
"url": "https://files.pythonhosted.org/packages/43/a5/46ee0f07167bcfdde7f36245025903cc0d8b00ed197f8929f73439472101/autoannotator-0.2.0-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "b4b6c2197bf171c943b87b23d7b9b6185e5c5c98a77056bbb4a83fddf5e88fd9",
"md5": "b276327e9b24095a9b9aceff8a8f2493",
"sha256": "6f3b08fb54741f514825ba72bf6f4ae2c7084d20a0a6099f29510bf590e4400b"
},
"downloads": -1,
"filename": "autoannotator-0.2.0.tar.gz",
"has_sig": false,
"md5_digest": "b276327e9b24095a9b9aceff8a8f2493",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.10",
"size": 1173086,
"upload_time": "2024-01-20T09:22:02",
"upload_time_iso_8601": "2024-01-20T09:22:02.908320Z",
"url": "https://files.pythonhosted.org/packages/b4/b6/c2197bf171c943b87b23d7b9b6185e5c5c98a77056bbb4a83fddf5e88fd9/autoannotator-0.2.0.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-01-20 09:22:02",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "CatsWhoTrain",
"github_project": "autoannotator",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"requirements": [
{
"name": "loguru",
"specs": []
},
{
"name": "onnx",
"specs": []
},
{
"name": "onnxruntime-gpu",
"specs": []
},
{
"name": "pydantic",
"specs": []
},
{
"name": "scikit-image",
"specs": []
},
{
"name": "scikit-learn",
"specs": []
},
{
"name": "opencv-python",
"specs": []
},
{
"name": "pytest",
"specs": []
},
{
"name": "requests",
"specs": []
},
{
"name": "tqdm",
"specs": []
}
],
"lcname": "autoannotator"
}