vl-datasets


Namevl-datasets JSON
Version 0.0.11 PyPI version JSON
download
home_pagehttps://github.com/visual-layer/vl-datasets
SummaryOpen, Clean Datasets for Computer Vision.
upload_time2023-05-24 09:38:50
maintainer
docs_urlNone
authorVisual Layer
requires_python>=3.7
licenseApache-2.0
keywords machine learning computer vision data-centric
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            
<!-- PROJECT SHIELDS -->
<!--
*** I'm using markdown "reference style" links for readability.
*** Reference links are enclosed in brackets [ ] instead of parentheses ( ).
*** See the bottom of this document for the declaration of the reference variables
*** for contributors-url, forks-url, etc. This is an optional, concise syntax you may use.
*** https://www.markdownguide.org/basic-syntax/#reference-style-links
-->

[![PyPi][pypi-shield]][pypi-url]
[![PyPi][pypiversion-shield]][pypi-url]
[![PyPi][downloads-shield]][downloads-url]
[![License][license-shield]][license-url]
[![TestedOn][testedon-shield]][pypi-url]
<!-- [![Contributors][contributors-shield]][contributors-url] -->


<!-- MARKDOWN LINKS & IMAGES -->
<!-- https://www.markdownguide.org/basic-syntax/#reference-style-links -->
[pypi-shield]: https://img.shields.io/badge/Python-3.7%20--%203.11-blue?style=for-the-badge
[pypi-url]: https://pypi.org/project/vl-datasets/
[pypiversion-shield]: https://img.shields.io/pypi/v/vl-datasets?style=for-the-badge
[downloads-shield]: https://img.shields.io/badge/dynamic/json?style=for-the-badge&label=downloads&query=%24.total_downloads&url=https%3A%2F%2Fapi.pepy.tech%2Fapi%2Fprojects%2Fvl-datasets&color=lightblue
[downloads-url]: https://pypi.org/project/vl-datasets/
<!-- [contributors-shield]: https://img.shields.io/github/contributors/visual-layer/fastdup?style=for-the-badge -->
<!-- [contributors-url]: https://github.com/othneildrew/Best-README-Template/graphs/contributors -->
[license-shield]: https://img.shields.io/badge/License-Apache%202.0-purple.svg?style=for-the-badge
[license-url]: https://github.com/visual-layer/vl-datasets/blob/main/LICENSE
[testedon-shield]: https://img.shields.io/badge/Tested%20on-Ubuntu--22.04%20%7C%20MacOS--10.16%20Intel%20%7C%20Windows%2010-brightgreen?style=for-the-badge


<!-- PROJECT LOGO -->
<br />
<div align="center">
<a href="https://www.visual-layer.com">
  <img alt="Visual Layer Logo" src="https://raw.githubusercontent.com/visual-layer/fastdup/main/gallery/visual_layer_logo.png" alt="Logo" width="350">
</a>
<h3 align="center">VL-Datasets</h3>
<h4 align="center">Open, Clean, Curated Datasets for Computer Vision</h4>

  <p align="center">
  <br />
    🔥 We use
    <a href="https://github.com/visual-layer/fastdup">fastdup</a> - a free tool to clean all datasets shared in this repo.
    <br />
    <a href="https://visual-layer.readme.io/" target="_blank" rel="noopener noreferrer"><strong>Explore the docs »</strong></a>
    <br />
    <a href="https://github.com/visual-layer/vl-datasets/issues" target="_blank" rel="noopener noreferrer">Report Issues</a>
    ·
    <a href="https://medium.com/@visual-layer/" target="_blank" rel="noopener noreferrer">Read Blog</a>
    ·
    <a href="mailto:info@visual-layer.com?subject=Sign-up%20for%20access" target="_blank" rel="noopener noreferrer">Get In Touch</a>
    ·
    <a href="https://visual-layer.com/" target="_blank" rel="noopener noreferrer">About Us</a>
    <br />
    <br /> 
    <a href="https://visualdatabase.slack.com/join/shared_invite/zt-19jaydbjn-lNDEDkgvSI1QwbTXSY6dlA#/shared-invite/email" target="_blank" rel="noopener noreferrer">
    <img src="https://img.shields.io/badge/JOIN US ON SLACK-4A154B?style=for-the-badge&logo=slack&logoColor=white" alt="Logo">
    </a>
    <a href="https://visual-layer.readme.io/discuss" target="_blank" rel="noopener noreferrer">
    <img src="https://img.shields.io/badge/DISCUSSION%20FORUM-brightgreen?style=for-the-badge&logo=discourse&logoWidth=20" alt="Logo">
    </a>
    <a href="https://www.linkedin.com/company/visual-layer/" target="_blank" rel="noopener noreferrer">
    <img src="https://img.shields.io/badge/LinkedIn-0077B5?style=for-the-badge&logo=linkedin&logoColor=white" alt="Logo">
    </a>
    <a href="https://twitter.com/visual_layer" target="_blank" rel="noopener noreferrer">
    <img src="https://img.shields.io/badge/Twitter-1DA1F2?style=for-the-badge&logo=twitter&logoColor=white" alt="Logo">
    </a>
    <a href="https://www.youtube.com/@visual-layer4035" target="_blank" rel="noopener noreferrer">
    <img src="https://img.shields.io/badge/-YouTube-black.svg?style=for-the-badge&logo=youtube&colorB=red" alt="Logo">
    </a>
  </p>
</div>

## Description

`vl-datasets` is a Python package that provides access to clean computer vision datasets with only 2 lines of code.

For example, to get access to the clean version of the [Food-101](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/) dataset simply run:

![image](./imgs/usage.png)

We support some of the most widely used computer vision datasets.
[Let us know](https://forms.gle/8jxPkyzeKj82kPed8) if you have additional request to support a new dataset.

All the datasets are analyzed for issues such as: 

+ Duplicates.
+ Near Duplicates.
+ Broken images.
+ Outliers.
+ Dark/Bright/Blurry images.
+ Mislabels.
+ Data Leakage.

![image](./imgs/issues.png)




## Why?

Computer vision is an exciting and rapidly advancing field, with new techniques and models emerging now and then. 
However, to develop and evaluate these models, it's essential to have reliable and standardized datasets to work with.

Even with the recent success of generative models, data quality remains an issue that's [mainly overlooked](https://medium.com/@amiralush/large-image-datasets-today-are-a-mess-e3ea4c9e8d22).
Training models will erroneours data impacts model accuracy, incurs costs in time, storage and computational resources.

We believe that access to clean and high-quality computer vision datasets leads to accurate, non-biased, and efficient model.
By providing public access to `vl-datasets` we hope it helps advance the field of computer vision.

## Datasets & Access

`vl-datasets` provides a convenient way to access the cleaned version of the datasets in Python.

Alternatively, for each dataset in this repo, we provide a `.csv` file that lists the problematic images from the dataset.

You can use the listed images in the `.csv` to improve the model by re-labeling the them or just simply remove it from the dataset.

We're a startup and we'd like to offer free access to the datasets as much as we can afford to. But in doing so, we'd also need your support.

We're offering select `.csv` files completely free with no strings attached. 
For access to our complete dataset and exclusive beta features, all we ask is that you [sign up](https://forms.gle/8jxPkyzeKj82kPed8) to be a beta tester – it's completely free and your feedback will help shape the future of our platform. 

Here is a table of widely used computer vision datasets, issues we found and a link to access the `.csv` file.

| Dataset                                                                 | Issues                                                                                                                                                                                 | CSV                                                                                                | Import Statement                |
|-------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------|----------------------|
| [Food-101](./dataset_card/food101/card.md) | <ul><li>Duplicates - 0.233 % (235)</li><li>Outliers - 0.076 % (77)</li><li>Blur - Blur - 0.183 % (185)</li><li>Dark - 0.043 % (43)</li><li><b>Total</b> - 0.535 % (540)</li></ul><div align="right"><a href="./dataset_card/food101/card.md"><strong>More »</strong></a></div>| Download [here](https://drive.google.com/uc?export=download&id=1ZG5GvU342l4YmSeYo6v6LeKbMM5fwjjw). | `from vl_datasets import VLFood101`       |
| [Oxford-IIIT Pet](./dataset_card/oxford-iiit-pets/card.md)          | <ul><li>Duplicates - 1.021% (75)</li><li>Outliers - 0.095% (7)</li><li>Dark - 0.054% (4)</li><li><b>Total</b> - 1.170 % (86)</li></ul><div align="right"><a href="./dataset_card/oxford-iiit-pets/card.md"><strong>More »</strong></a></div>         | Download [here](https://drive.google.com/uc?export=download&id=1OLa8k4NITnmCHjeByzvGaWt3W7k6R1QL). | `from vl_datasets import VLOxfordIIITPet` |
| [LAION-1B](./dataset_card/laion-1b/card.md)                            | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul><div align="right"><a href="./dataset_card/laion-1b/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [ImageNet-21K](./dataset_card/imagenet-1k/card.md)                              | <ul><li>Duplicates - 11.853 % (1,559,120)</li><li>Outliers - 0.085 % (11,119)</li><li>Blur - 0.292 % (38,458)</li><li>Dark - 0.179 % (23,574)</li><li>Bright - 0.431 % (56,754)</li><li>Mislabels - 3.064 % (402,963)</li><li><b>Total</b> - 15.904 % (2,091,988)</li></ul><div align="right"><a href="./dataset_card/imagenet-1k/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [ImageNet-1K](./dataset_card/imagenet-21k/card.md)                                | <ul><li>Duplicates - 0.520 % (6,660)</li><li>Outliers - 0.090 % (1,150)</li><li>Blur - 0.200 % (2,554)</li><li>Dark - 0.244 % (2,997)</li><li>Bright - 0.058 % (746)</li><li>Mislabels - 0.119 % (1,518)</li><li><b>Total</b> - 1.221 % (15,625)</li></ul><div align="right"><a href="./dataset_card/imagenet-21k/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [KITTI](./dataset_card/kitti/card.md)                          | <ul><li>Duplicates - 15.294 % (2294)</li><li>Outliers - 0.107 % (16)</li><li><b>Total</b> - 15.401 % (2310)</li></ul><div align="right"><a href="./dataset_card/kitti/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [DeepFashion](./dataset_card/deep-fashion/card.md)     | <ul><li>Duplicates - 5.114 % (14,772)</li><li>Outliers - 0.037 % (107)</li><b>Total</b> - 5.151 % (14,879)</li></ul><div align="right"><a href="./dataset_card/deep-fashion/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [CelebA-HQ](./dataset_card/celeb-a-hq/card.md)           | <ul><li>Duplicates - 1.673 % (3,389)</li><li>Outliers - 0.077 % (157)</li><li>Blur - 0.512 % (1,037)</li><li>Dark - 0.009 % (18)</li><li>Mislabels - 0.006 % (13)</li><li><b>Total</b> - 2.277 % (4,614)</li></ul><div align="right"><a href="./dataset_card/celeb-a-hq/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [COCO](./dataset_card/coco/card.md)                                   | <ul><li>Duplicates - 0.123 % (201)</li><li>Outliers - 0.087 % (143)</li><li>Blur - 0.029 % (47)</li><li>Dark - 0.106 % (174)</li><li>Bright - 0.013 % (21)</li><li><b>Total</b> - 0.358 % (586)</li></ul><div align="right"><a href="./dataset_card/coco/card.md"><strong>More »</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
<!-- | [ADE20K](https://groups.csail.mit.edu/vision/datasets/ADE20K/)          | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
| [Places365](https://github.com/CSAILVision/places365)                   | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |
 -->


Learn more on how we clean the datasets using our profilling tool [here](https://visual-layer.link).


## Installation

**Option 1** - Install `vl_datasets` package from [PyPI](https://pypi.org/project/vl-datasets/):

```shell
pip install vl-datasets
```

**Option 2** - Install the bleeding edge version on GitHub:
```
pip install git+https://github.com/visual-layer/vl-datasets.git@main --upgrade
```

## Usage
To start using `vl-datasets`, import the clean version of the dataset with:

```python
from vl_datasets import VLFood101
```

This should import the clean version of the `Food101` dataset.

Next, you can load the dataset as a PyTorch `Dataset`.

```python
train_dataset = VLFood101('./', split='train')
valid_dataset = VLFood101('./', split='test')
```

If you have a custom `.csv` file you can optionally pass in the file:

```python
train_dataset = VLFood101('./', split='train', exclude_csv='my-file.csv')
```
The filenames listed in the `.csv` will be excluded in the dataset.

Next, you can load the train and validation datasets in a PyTorch training loop.

See the [Learn from Examples](#learn-from-examples) section to learn more.


> **NOTE**: Sign up [here](https://forms.gle/8jxPkyzeKj82kPed8) for free to be our beta testers and get full access to the all the `.csv` files for the dataset listed in this repo. 

With the dataset loaded you can train a model using PyTorch training loop.

## Learn from Examples

<table>
	<tr>
		<td rowspan="4" width="160">
			<a href="https://visual-layer.readme.io/docs/getting-started">
				<img src="./imgs/food.jpg" width="256" />
			</a>
		</td>
		<td rowspan="4">
			<ul>
				<li><b>Dataset:</b> <code>VLFood101</code></li>
				<li><b>Framework:</b> PyTorch.</li>
				<li><b>Description:</b> Load a dataset and train a PyTorch model.</li>
			</ul>
		</td>
		<td align="center" width="80">
			<a href="https://nbviewer.org/github/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb">
				<img src="./imgs/nbviewer_logo.svg" height="34" />
			</a>
		</td>
	</tr>
	<tr>
		<td align="center">
			<a href="https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb">
				<img src="./imgs/github_logo.png" height="32" />
			</a>
		</td>
	</tr>
	<tr>
		<td align="center">
			<a href="https://colab.research.google.com/github/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb">
				<img src="./imgs/colab_logo.png" height="28" />
			</a>
		</td>
	</tr>
    <tr>
		<td align="center">
			<a href="https://kaggle.com/kernels/welcome?src=https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb">
				<img src="./imgs/kaggle_logo.png" height="28" />
			</a>
		</td>
	</tr>
	<!-- ------------------------------------------------------------------- -->
	<tr>
		<td rowspan="4" width="160">
			<a href="https://visual-layer.readme.io/docs/objects-and-bounding-boxes">
				<img src="./imgs/pet.jpg" width="256" />
			</a>
		</td>
		<td rowspan="4">
			<ul>
				<li><b>Dataset:</b> <code>VLOxfordIIITPet</code></li>
				<li><b>Framework:</b> fast.ai.</li>
				<li><b>Description:</b> Finetune a pretrained TIMM model using fastai.</li>
			</ul>
		</td>
		<td align="center" width="80">
			<a href="https://nbviewer.org/github/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb">
				<img src="./imgs/nbviewer_logo.svg" height="34" />
			</a>
		</td>
	</tr>
	<tr>
		<td align="center">
			<a href="https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb">
				<img src="./imgs/github_logo.png" height="32" />
			</a>
		</td>
	</tr>
	<tr>
		<td align="center">
			<a href="https://colab.research.google.com/github/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb">
				<img src="./imgs/colab_logo.png" height="28" />
			</a>
		</td>
	</tr>
    <tr>
		<td align="center">
			<a href="https://kaggle.com/kernels/welcome?src=https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb">
				<img src="./imgs/kaggle_logo.png" height="28" />
			</a>
		</td>
	</tr>
	<!-- ------------------------------------------------------------------- -->
</table>


## License
`vl-datasets` is licensed under the Apache 2.0 License. See [LICENSE](./LICENSE).

However, you are bound to the usage license of the original dataset. It is your responsibility to determine whether you have permission to use the dataset under the dataset's license. We provide no warranty or guarantee of accuracy or completeness.

## Usage Tracking
This repository incorporates usage tracking using [Sentry.io](https://sentry.io/) to monitor and collect valuable information about the usage of the application.

Usage tracking allows us to gain insights into how the application is being used in real-world scenarios. It provides us with valuable information that helps in understanding user behavior, identifying potential issues, and making informed decisions to improve the application.

We DO NOT collect folder names, user names, image names, image content and other personaly identifiable information.

What data is tracked?
+ **Errors and Exceptions**: Sentry captures errors and exceptions that occur in the application, providing detailed stack traces and relevant information to help diagnose and fix issues.
+ **Performance Metrics**: Sentry collects performance metrics, such as response times, latency, and resource usage, enabling us to monitor and optimize the application's performance.

To opt out, define an environment variable named `SENTRY_OPT_OUT`. 

On Linux run the following:
```bash
export SENTRY_OPT_OUT=True
```

Read more on Sentry's official [webpage](https://sentry.io/welcome/).


## Getting Help
Get help from the Visual Layer team or community members via the following channels -
+ [Slack](https://visualdatabase.slack.com/join/shared_invite/zt-19jaydbjn-lNDEDkgvSI1QwbTXSY6dlA#/shared-invite/email).
+ GitHub [issues](https://github.com/visual-layer/vl-datasets/issues).
+ Discussion [forum](https://visual-layer.readme.io/discuss).

## About Visual-Layer

<div align="center">
<a href="https://www.visual-layer.com">
  <img alt="Visual Layer Logo" src="https://raw.githubusercontent.com/visual-layer/fastdup/main/gallery/visual_layer_logo.png" alt="Logo" width="250">
</a>
</div>


Visual Layer is founded by the authors of [XGBoost](https://github.com/apache/tvm), [Apache TVM](https://github.com/apache/tvm) & [Turi Create](https://github.com/apple/turicreate) - [Danny Bickson](https://www.linkedin.com/in/dr-danny-bickson-835b32), [Carlos Guestrin](https://www.linkedin.com/in/carlos-guestrin-5352a869) and [Amir Alush](https://www.linkedin.com/in/amiralush).

Learn more about Visual Layer [here](https://visual-layer.com).

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/visual-layer/vl-datasets",
    "name": "vl-datasets",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.7",
    "maintainer_email": "",
    "keywords": "machine learning,computer vision,data-centric",
    "author": "Visual Layer",
    "author_email": "info@visual-layer.com",
    "download_url": "",
    "platform": null,
    "description": "\n<!-- PROJECT SHIELDS -->\n<!--\n*** I'm using markdown \"reference style\" links for readability.\n*** Reference links are enclosed in brackets [ ] instead of parentheses ( ).\n*** See the bottom of this document for the declaration of the reference variables\n*** for contributors-url, forks-url, etc. This is an optional, concise syntax you may use.\n*** https://www.markdownguide.org/basic-syntax/#reference-style-links\n-->\n\n[![PyPi][pypi-shield]][pypi-url]\n[![PyPi][pypiversion-shield]][pypi-url]\n[![PyPi][downloads-shield]][downloads-url]\n[![License][license-shield]][license-url]\n[![TestedOn][testedon-shield]][pypi-url]\n<!-- [![Contributors][contributors-shield]][contributors-url] -->\n\n\n<!-- MARKDOWN LINKS & IMAGES -->\n<!-- https://www.markdownguide.org/basic-syntax/#reference-style-links -->\n[pypi-shield]: https://img.shields.io/badge/Python-3.7%20--%203.11-blue?style=for-the-badge\n[pypi-url]: https://pypi.org/project/vl-datasets/\n[pypiversion-shield]: https://img.shields.io/pypi/v/vl-datasets?style=for-the-badge\n[downloads-shield]: https://img.shields.io/badge/dynamic/json?style=for-the-badge&label=downloads&query=%24.total_downloads&url=https%3A%2F%2Fapi.pepy.tech%2Fapi%2Fprojects%2Fvl-datasets&color=lightblue\n[downloads-url]: https://pypi.org/project/vl-datasets/\n<!-- [contributors-shield]: https://img.shields.io/github/contributors/visual-layer/fastdup?style=for-the-badge -->\n<!-- [contributors-url]: https://github.com/othneildrew/Best-README-Template/graphs/contributors -->\n[license-shield]: https://img.shields.io/badge/License-Apache%202.0-purple.svg?style=for-the-badge\n[license-url]: https://github.com/visual-layer/vl-datasets/blob/main/LICENSE\n[testedon-shield]: https://img.shields.io/badge/Tested%20on-Ubuntu--22.04%20%7C%20MacOS--10.16%20Intel%20%7C%20Windows%2010-brightgreen?style=for-the-badge\n\n\n<!-- PROJECT LOGO -->\n<br />\n<div align=\"center\">\n<a href=\"https://www.visual-layer.com\">\n  <img alt=\"Visual Layer Logo\" src=\"https://raw.githubusercontent.com/visual-layer/fastdup/main/gallery/visual_layer_logo.png\" alt=\"Logo\" width=\"350\">\n</a>\n<h3 align=\"center\">VL-Datasets</h3>\n<h4 align=\"center\">Open, Clean, Curated Datasets for Computer Vision</h4>\n\n  <p align=\"center\">\n  <br />\n    \ud83d\udd25 We use\n    <a href=\"https://github.com/visual-layer/fastdup\">fastdup</a> - a free tool to clean all datasets shared in this repo.\n    <br />\n    <a href=\"https://visual-layer.readme.io/\" target=\"_blank\" rel=\"noopener noreferrer\"><strong>Explore the docs \u00bb</strong></a>\n    <br />\n    <a href=\"https://github.com/visual-layer/vl-datasets/issues\" target=\"_blank\" rel=\"noopener noreferrer\">Report Issues</a>\n    \u00b7\n    <a href=\"https://medium.com/@visual-layer/\" target=\"_blank\" rel=\"noopener noreferrer\">Read Blog</a>\n    \u00b7\n    <a href=\"mailto:info@visual-layer.com?subject=Sign-up%20for%20access\" target=\"_blank\" rel=\"noopener noreferrer\">Get In Touch</a>\n    \u00b7\n    <a href=\"https://visual-layer.com/\" target=\"_blank\" rel=\"noopener noreferrer\">About Us</a>\n    <br />\n    <br /> \n    <a href=\"https://visualdatabase.slack.com/join/shared_invite/zt-19jaydbjn-lNDEDkgvSI1QwbTXSY6dlA#/shared-invite/email\" target=\"_blank\" rel=\"noopener noreferrer\">\n    <img src=\"https://img.shields.io/badge/JOIN US ON SLACK-4A154B?style=for-the-badge&logo=slack&logoColor=white\" alt=\"Logo\">\n    </a>\n    <a href=\"https://visual-layer.readme.io/discuss\" target=\"_blank\" rel=\"noopener noreferrer\">\n    <img src=\"https://img.shields.io/badge/DISCUSSION%20FORUM-brightgreen?style=for-the-badge&logo=discourse&logoWidth=20\" alt=\"Logo\">\n    </a>\n    <a href=\"https://www.linkedin.com/company/visual-layer/\" target=\"_blank\" rel=\"noopener noreferrer\">\n    <img src=\"https://img.shields.io/badge/LinkedIn-0077B5?style=for-the-badge&logo=linkedin&logoColor=white\" alt=\"Logo\">\n    </a>\n    <a href=\"https://twitter.com/visual_layer\" target=\"_blank\" rel=\"noopener noreferrer\">\n    <img src=\"https://img.shields.io/badge/Twitter-1DA1F2?style=for-the-badge&logo=twitter&logoColor=white\" alt=\"Logo\">\n    </a>\n    <a href=\"https://www.youtube.com/@visual-layer4035\" target=\"_blank\" rel=\"noopener noreferrer\">\n    <img src=\"https://img.shields.io/badge/-YouTube-black.svg?style=for-the-badge&logo=youtube&colorB=red\" alt=\"Logo\">\n    </a>\n  </p>\n</div>\n\n## Description\n\n`vl-datasets` is a Python package that provides access to clean computer vision datasets with only 2 lines of code.\n\nFor example, to get access to the clean version of the [Food-101](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/) dataset simply run:\n\n![image](./imgs/usage.png)\n\nWe support some of the most widely used computer vision datasets.\n[Let us know](https://forms.gle/8jxPkyzeKj82kPed8) if you have additional request to support a new dataset.\n\nAll the datasets are analyzed for issues such as: \n\n+ Duplicates.\n+ Near Duplicates.\n+ Broken images.\n+ Outliers.\n+ Dark/Bright/Blurry images.\n+ Mislabels.\n+ Data Leakage.\n\n![image](./imgs/issues.png)\n\n\n\n\n## Why?\n\nComputer vision is an exciting and rapidly advancing field, with new techniques and models emerging now and then. \nHowever, to develop and evaluate these models, it's essential to have reliable and standardized datasets to work with.\n\nEven with the recent success of generative models, data quality remains an issue that's [mainly overlooked](https://medium.com/@amiralush/large-image-datasets-today-are-a-mess-e3ea4c9e8d22).\nTraining models will erroneours data impacts model accuracy, incurs costs in time, storage and computational resources.\n\nWe believe that access to clean and high-quality computer vision datasets leads to accurate, non-biased, and efficient model.\nBy providing public access to `vl-datasets` we hope it helps advance the field of computer vision.\n\n## Datasets & Access\n\n`vl-datasets` provides a convenient way to access the cleaned version of the datasets in Python.\n\nAlternatively, for each dataset in this repo, we provide a `.csv` file that lists the problematic images from the dataset.\n\nYou can use the listed images in the `.csv` to improve the model by re-labeling the them or just simply remove it from the dataset.\n\nWe're a startup and we'd like to offer free access to the datasets as much as we can afford to. But in doing so, we'd also need your support.\n\nWe're offering select `.csv` files completely free with no strings attached. \nFor access to our complete dataset and exclusive beta features, all we ask is that you [sign up](https://forms.gle/8jxPkyzeKj82kPed8) to be a beta tester \u2013 it's completely free and your feedback will help shape the future of our platform. \n\nHere is a table of widely used computer vision datasets, issues we found and a link to access the `.csv` file.\n\n| Dataset                                                                 | Issues                                                                                                                                                                                 | CSV                                                                                                | Import Statement                |\n|-------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------|----------------------|\n| [Food-101](./dataset_card/food101/card.md) | <ul><li>Duplicates - 0.233 % (235)</li><li>Outliers - 0.076 % (77)</li><li>Blur - Blur - 0.183 % (185)</li><li>Dark - 0.043 % (43)</li><li><b>Total</b> - 0.535 % (540)</li></ul><div align=\"right\"><a href=\"./dataset_card/food101/card.md\"><strong>More \u00bb</strong></a></div>| Download [here](https://drive.google.com/uc?export=download&id=1ZG5GvU342l4YmSeYo6v6LeKbMM5fwjjw). | `from vl_datasets import VLFood101`       |\n| [Oxford-IIIT Pet](./dataset_card/oxford-iiit-pets/card.md)          | <ul><li>Duplicates - 1.021% (75)</li><li>Outliers - 0.095% (7)</li><li>Dark - 0.054% (4)</li><li><b>Total</b> - 1.170 % (86)</li></ul><div align=\"right\"><a href=\"./dataset_card/oxford-iiit-pets/card.md\"><strong>More \u00bb</strong></a></div>         | Download [here](https://drive.google.com/uc?export=download&id=1OLa8k4NITnmCHjeByzvGaWt3W7k6R1QL). | `from vl_datasets import VLOxfordIIITPet` |\n| [LAION-1B](./dataset_card/laion-1b/card.md)                            | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul><div align=\"right\"><a href=\"./dataset_card/laion-1b/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [ImageNet-21K](./dataset_card/imagenet-1k/card.md)                              | <ul><li>Duplicates - 11.853 % (1,559,120)</li><li>Outliers - 0.085 % (11,119)</li><li>Blur - 0.292 % (38,458)</li><li>Dark - 0.179 % (23,574)</li><li>Bright - 0.431 % (56,754)</li><li>Mislabels - 3.064 % (402,963)</li><li><b>Total</b> - 15.904 % (2,091,988)</li></ul><div align=\"right\"><a href=\"./dataset_card/imagenet-1k/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [ImageNet-1K](./dataset_card/imagenet-21k/card.md)                                | <ul><li>Duplicates - 0.520 % (6,660)</li><li>Outliers - 0.090 % (1,150)</li><li>Blur - 0.200 % (2,554)</li><li>Dark - 0.244 % (2,997)</li><li>Bright - 0.058 % (746)</li><li>Mislabels - 0.119 % (1,518)</li><li><b>Total</b> - 1.221 % (15,625)</li></ul><div align=\"right\"><a href=\"./dataset_card/imagenet-21k/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [KITTI](./dataset_card/kitti/card.md)                          | <ul><li>Duplicates - 15.294 % (2294)</li><li>Outliers - 0.107 % (16)</li><li><b>Total</b> - 15.401 % (2310)</li></ul><div align=\"right\"><a href=\"./dataset_card/kitti/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [DeepFashion](./dataset_card/deep-fashion/card.md)     | <ul><li>Duplicates - 5.114 % (14,772)</li><li>Outliers - 0.037 % (107)</li><b>Total</b> - 5.151 % (14,879)</li></ul><div align=\"right\"><a href=\"./dataset_card/deep-fashion/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [CelebA-HQ](./dataset_card/celeb-a-hq/card.md)           | <ul><li>Duplicates - 1.673 % (3,389)</li><li>Outliers - 0.077 % (157)</li><li>Blur - 0.512 % (1,037)</li><li>Dark - 0.009 % (18)</li><li>Mislabels - 0.006 % (13)</li><li><b>Total</b> - 2.277 % (4,614)</li></ul><div align=\"right\"><a href=\"./dataset_card/celeb-a-hq/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [COCO](./dataset_card/coco/card.md)                                   | <ul><li>Duplicates - 0.123 % (201)</li><li>Outliers - 0.087 % (143)</li><li>Blur - 0.029 % (47)</li><li>Dark - 0.106 % (174)</li><li>Bright - 0.013 % (21)</li><li><b>Total</b> - 0.358 % (586)</li></ul><div align=\"right\"><a href=\"./dataset_card/coco/card.md\"><strong>More \u00bb</strong></a></div> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n<!-- | [ADE20K](https://groups.csail.mit.edu/vision/datasets/ADE20K/)          | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n| [Places365](https://github.com/CSAILVision/places365)                   | <ul><li>Duplicates - WIP % (WIP)</li><li>Outliers - WIP % (WIP)</li><li>Broken - WIP % (WIP)</li><li>Blur - WIP % (WIP)</li><li>Dark - WIP % (WIP)</li><li>Bright - WIP % (WIP)</li></ul> | Request access [here](https://forms.gle/8jxPkyzeKj82kPed8).                                        | WIP                  |\n -->\n\n\nLearn more on how we clean the datasets using our profilling tool [here](https://visual-layer.link).\n\n\n## Installation\n\n**Option 1** - Install `vl_datasets` package from [PyPI](https://pypi.org/project/vl-datasets/):\n\n```shell\npip install vl-datasets\n```\n\n**Option 2** - Install the bleeding edge version on GitHub:\n```\npip install git+https://github.com/visual-layer/vl-datasets.git@main --upgrade\n```\n\n## Usage\nTo start using `vl-datasets`, import the clean version of the dataset with:\n\n```python\nfrom vl_datasets import VLFood101\n```\n\nThis should import the clean version of the `Food101` dataset.\n\nNext, you can load the dataset as a PyTorch `Dataset`.\n\n```python\ntrain_dataset = VLFood101('./', split='train')\nvalid_dataset = VLFood101('./', split='test')\n```\n\nIf you have a custom `.csv` file you can optionally pass in the file:\n\n```python\ntrain_dataset = VLFood101('./', split='train', exclude_csv='my-file.csv')\n```\nThe filenames listed in the `.csv` will be excluded in the dataset.\n\nNext, you can load the train and validation datasets in a PyTorch training loop.\n\nSee the [Learn from Examples](#learn-from-examples) section to learn more.\n\n\n> **NOTE**: Sign up [here](https://forms.gle/8jxPkyzeKj82kPed8) for free to be our beta testers and get full access to the all the `.csv` files for the dataset listed in this repo. \n\nWith the dataset loaded you can train a model using PyTorch training loop.\n\n## Learn from Examples\n\n<table>\n\t<tr>\n\t\t<td rowspan=\"4\" width=\"160\">\n\t\t\t<a href=\"https://visual-layer.readme.io/docs/getting-started\">\n\t\t\t\t<img src=\"./imgs/food.jpg\" width=\"256\" />\n\t\t\t</a>\n\t\t</td>\n\t\t<td rowspan=\"4\">\n\t\t\t<ul>\n\t\t\t\t<li><b>Dataset:</b> <code>VLFood101</code></li>\n\t\t\t\t<li><b>Framework:</b> PyTorch.</li>\n\t\t\t\t<li><b>Description:</b> Load a dataset and train a PyTorch model.</li>\n\t\t\t</ul>\n\t\t</td>\n\t\t<td align=\"center\" width=\"80\">\n\t\t\t<a href=\"https://nbviewer.org/github/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb\">\n\t\t\t\t<img src=\"./imgs/nbviewer_logo.svg\" height=\"34\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb\">\n\t\t\t\t<img src=\"./imgs/github_logo.png\" height=\"32\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://colab.research.google.com/github/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb\">\n\t\t\t\t<img src=\"./imgs/colab_logo.png\" height=\"28\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n    <tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://kaggle.com/kernels/welcome?src=https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-pytorch.ipynb\">\n\t\t\t\t<img src=\"./imgs/kaggle_logo.png\" height=\"28\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<!-- ------------------------------------------------------------------- -->\n\t<tr>\n\t\t<td rowspan=\"4\" width=\"160\">\n\t\t\t<a href=\"https://visual-layer.readme.io/docs/objects-and-bounding-boxes\">\n\t\t\t\t<img src=\"./imgs/pet.jpg\" width=\"256\" />\n\t\t\t</a>\n\t\t</td>\n\t\t<td rowspan=\"4\">\n\t\t\t<ul>\n\t\t\t\t<li><b>Dataset:</b> <code>VLOxfordIIITPet</code></li>\n\t\t\t\t<li><b>Framework:</b> fast.ai.</li>\n\t\t\t\t<li><b>Description:</b> Finetune a pretrained TIMM model using fastai.</li>\n\t\t\t</ul>\n\t\t</td>\n\t\t<td align=\"center\" width=\"80\">\n\t\t\t<a href=\"https://nbviewer.org/github/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb\">\n\t\t\t\t<img src=\"./imgs/nbviewer_logo.svg\" height=\"34\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb\">\n\t\t\t\t<img src=\"./imgs/github_logo.png\" height=\"32\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://colab.research.google.com/github/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb\">\n\t\t\t\t<img src=\"./imgs/colab_logo.png\" height=\"28\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n    <tr>\n\t\t<td align=\"center\">\n\t\t\t<a href=\"https://kaggle.com/kernels/welcome?src=https://github.com/visual-layer/vl-datasets/blob/main/notebooks/train-fastai.ipynb\">\n\t\t\t\t<img src=\"./imgs/kaggle_logo.png\" height=\"28\" />\n\t\t\t</a>\n\t\t</td>\n\t</tr>\n\t<!-- ------------------------------------------------------------------- -->\n</table>\n\n\n## License\n`vl-datasets` is licensed under the Apache 2.0 License. See [LICENSE](./LICENSE).\n\nHowever, you are bound to the usage license of the original dataset. It is your responsibility to determine whether you have permission to use the dataset under the dataset's license. We provide no warranty or guarantee of accuracy or completeness.\n\n## Usage Tracking\nThis repository incorporates usage tracking using [Sentry.io](https://sentry.io/) to monitor and collect valuable information about the usage of the application.\n\nUsage tracking allows us to gain insights into how the application is being used in real-world scenarios. It provides us with valuable information that helps in understanding user behavior, identifying potential issues, and making informed decisions to improve the application.\n\nWe DO NOT collect folder names, user names, image names, image content and other personaly identifiable information.\n\nWhat data is tracked?\n+ **Errors and Exceptions**: Sentry captures errors and exceptions that occur in the application, providing detailed stack traces and relevant information to help diagnose and fix issues.\n+ **Performance Metrics**: Sentry collects performance metrics, such as response times, latency, and resource usage, enabling us to monitor and optimize the application's performance.\n\nTo opt out, define an environment variable named `SENTRY_OPT_OUT`. \n\nOn Linux run the following:\n```bash\nexport SENTRY_OPT_OUT=True\n```\n\nRead more on Sentry's official [webpage](https://sentry.io/welcome/).\n\n\n## Getting Help\nGet help from the Visual Layer team or community members via the following channels -\n+ [Slack](https://visualdatabase.slack.com/join/shared_invite/zt-19jaydbjn-lNDEDkgvSI1QwbTXSY6dlA#/shared-invite/email).\n+ GitHub [issues](https://github.com/visual-layer/vl-datasets/issues).\n+ Discussion [forum](https://visual-layer.readme.io/discuss).\n\n## About Visual-Layer\n\n<div align=\"center\">\n<a href=\"https://www.visual-layer.com\">\n  <img alt=\"Visual Layer Logo\" src=\"https://raw.githubusercontent.com/visual-layer/fastdup/main/gallery/visual_layer_logo.png\" alt=\"Logo\" width=\"250\">\n</a>\n</div>\n\n\nVisual Layer is founded by the authors of [XGBoost](https://github.com/apache/tvm), [Apache TVM](https://github.com/apache/tvm) & [Turi Create](https://github.com/apple/turicreate) - [Danny Bickson](https://www.linkedin.com/in/dr-danny-bickson-835b32), [Carlos Guestrin](https://www.linkedin.com/in/carlos-guestrin-5352a869) and [Amir Alush](https://www.linkedin.com/in/amiralush).\n\nLearn more about Visual Layer [here](https://visual-layer.com).\n",
    "bugtrack_url": null,
    "license": "Apache-2.0",
    "summary": "Open, Clean Datasets for Computer Vision.",
    "version": "0.0.11",
    "project_urls": {
        "Homepage": "https://github.com/visual-layer/vl-datasets"
    },
    "split_keywords": [
        "machine learning",
        "computer vision",
        "data-centric"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "e425594b740c0e07f40e8e2c2de9ebcbdd2172b0cc4364c91f06585da1313b52",
                "md5": "c94aaf23535bba7d6208c9bca76e570b",
                "sha256": "a22ce352c1ee762f6744c7bbd32af34101b9d5c3935039c4cffc77664d392e66"
            },
            "downloads": -1,
            "filename": "vl_datasets-0.0.11-py3.10-none-any.whl",
            "has_sig": false,
            "md5_digest": "c94aaf23535bba7d6208c9bca76e570b",
            "packagetype": "bdist_wheel",
            "python_version": "py3.10",
            "requires_python": ">=3.7",
            "size": 17233,
            "upload_time": "2023-05-24T09:38:50",
            "upload_time_iso_8601": "2023-05-24T09:38:50.728311Z",
            "url": "https://files.pythonhosted.org/packages/e4/25/594b740c0e07f40e8e2c2de9ebcbdd2172b0cc4364c91f06585da1313b52/vl_datasets-0.0.11-py3.10-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "9db56842e379667b8c8dcfef0aaa0e51cc3d046fccda81d8d3fd77d312ed9f24",
                "md5": "82766dd80de0c015e26c226b4e38a7bc",
                "sha256": "c5c441a4f9eba007f78505a9eb2a83a2ff1a53144da7f32457db2141bc479d8b"
            },
            "downloads": -1,
            "filename": "vl_datasets-0.0.11-py3.9-none-any.whl",
            "has_sig": false,
            "md5_digest": "82766dd80de0c015e26c226b4e38a7bc",
            "packagetype": "bdist_wheel",
            "python_version": "py3.9",
            "requires_python": ">=3.7",
            "size": 17247,
            "upload_time": "2023-05-24T09:38:52",
            "upload_time_iso_8601": "2023-05-24T09:38:52.487835Z",
            "url": "https://files.pythonhosted.org/packages/9d/b5/6842e379667b8c8dcfef0aaa0e51cc3d046fccda81d8d3fd77d312ed9f24/vl_datasets-0.0.11-py3.9-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2023-05-24 09:38:50",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "visual-layer",
    "github_project": "vl-datasets",
    "github_not_found": true,
    "lcname": "vl-datasets"
}
        
Elapsed time: 0.06742s