duplicate_images


Nameduplicate_images JSON
Version 0.11.0 PyPI version JSON
download
home_pagehttps://github.com/lene/DuplicateImages
SummaryFinds equal or similar images in a directory containing (many) image files
upload_time2024-01-25 21:54:35
maintainer
docs_urlNone
authorLene Preuss
requires_python>=3.9
license
keywords
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Finding Duplicate Images

Finds equal or similar images in a directory containing (many) image files.

Official home page: https://github.com/lene/DuplicateImages

Development page: https://gitlab.com/lilacashes/DuplicateImages

PyPI page: https://pypi.org/project/duplicate-images

## Usage

Installing:
```shell
$ pip install duplicate_images
```

Printing the help screen:
```shell
$ find-dups -h
```

Quick test run:
```shell
$ find-dups $IMAGE_ROOT 
```

Typical usage:
```shell
$ find-dups $IMAGE_ROOT --parallel --progress --hash-db hashes.json
```

### Supported image formats

* JPEG and PNG (tested quite thoroughly)
* HEIC (experimental support, tested cursorily only)
* All other 
  [formats supported](https://pillow.readthedocs.io/en/latest/handbook/image-file-formats.html) by 
  the `pillow` Python Imaging Library should work, but are not specifically tested. 

#### Explicitly allow huge images

The `PIL` image library, which is used as backend, limits the size of images to 178956970 pixels by
default, to guard against memory exhaustion. For larger images, specify the maximum image size in 
pixels with the `--max-image-pixels` option.

### Image comparison algorithms

Use the `--algorithm` option to select how equal images are found. The default algorithm is `phash`.

`ahash`, `colorhash`, `dhash`, `dhash_vertical`, `phash`, `phash_simple`, `whash`: seven different 
image hashing algorithms. See https://pypi.org/project/ImageHash for an introduction on image 
hashing and https://tech.okcupid.com/evaluating-perceptual-image-hashes-okcupid for some gory 
details which image hashing algorithm performs best in which situation. For a start I recommend 
using `phash`, and only evaluating the other algorithms if `phash` does not perform satisfactorily 
in your use case.

### Image similarity threshold configuration

Use the `--hash-size` parameter to tune the precision of the hashing algorithms. For the `colorhash`
algorithm the hash size is interpreted as the number of bin bits and defaults to 3. For all other
algorithms the hash size defaults to 8. For `whash` it must be a power of 2.

Use the `--max-distance` parameter to tune how close images should be to be considered duplicates.
The argument is a positive integer. Its value is highly dependent on the algorithm used and the 
nature of the images compared, so the best value for your use case can oly be found through 
experimentation.

**NOTE:** using the `--max-distance` parameter slows down the comparison considerably with large
image collections, making the runtime complexity go from O(N) to O(N<sup>2</sup>). If you want to 
scan collections with at least thousands of images, it is highly recommended to tune the desired 
similarity threshold with the `--hash-size` parameter alone, if that is at all possible.

### Pre-storing and using image hashes to speed up computation

Use the `--hash-db ${FILE}.json` or `--hash-db ${FILE}.pickle` option to store image hashes in the 
file `$FILE` in JSON or Pickle format and read image hashes from that file if they are already 
present there. This avoids having to compute the image hashes anew at every run and can 
significantly speed up run times.

### Handling matching images either as pairs or as groups

By default, matching images are presented as pairs. With the `--group` CLI option, they are handled
as a group containing all matching images.

Example: `1.jpg`, `2.jpg` and `3.jpg` in the current folder `.` are equal.

```shell
$ find-dups .
1.jpg 2.jpg
1.jpg 3.jpg
2.jpg 3.jpg
$ find-dups . --group
1.jpg 2.jpg 3.jpg
```

### Actions for matching image groups

Use the `--on-equal` option to select what to do to pairs of equal images. The default action is 
`print`.
- `delete-first` or `d1`: deletes the first of the files in the group
- `delete-last` or `dl`: deletes the last of the files in the group
- `delete-biggest` or `d>`: deletes the file with the biggest size
- `delete-smallest` or `d<`: deletes the file with the smallest size
- `symlink-smaller`: delete the smaller files and replace them to a symlink to the biggest file
- `eog`: launches the `eog` image viewer to compare the files in the group (*deprecated* by `exec`)
- `xv`: launches the `xv` image viewer to compare the files in the group (*deprecated* by `exec`)
- `print`: prints the files in the group
- `print_inline`: like `print` but without newline
- `quote`: prints the files in the group quoted for POSIX shells
- `quote_inline`: like `quote` but without newline
- `exec`: executes a command (see `--exec` argument below)
- `none`: does nothing; may be useful for benchmarking and testing

The `--exec` argument allows calling another program when the `--on-equal exec` option is given.
You can pass a command line string like `--exec "program {1} {2}"` where `{1}` and `{2}` are
replaced by the matching pair files (or first two files in a group), quoted so the shell recognizes
the files properly. The wildcard `{*}` expands to all files in a matching group, which when called
with the `--group` argument may be more than two images considered equal.

#### Examples:
* `--exec "open -a Preview -W {1} {2}"`: Opens the files in MacOS Preview app and waits for it.
* `--exec "ls -s {*}"`: Prints the size (in blocks) next to all files.
* `--exec 'for i in {*}; do dirname $i; basename $i; done'`: Shows the directory and the filename
  separately for all files.

### Parallel execution

Use the `--parallel` option to utilize all free cores on your system for calculating image hashes.
Optionally, you can specify the number of processes to use with `--parallel $N`.

To execute the `--on-equal` actions in parallel, use the `--parallel-actions` option, which also can
take an optional number of processes to use as argument.

### Excluding subfolders

Use the `--exclude-dir` option to exclude subfolders of `$IMAGE_ROOT` from the search. The argument
is a regular expression matching the subfolder names to be excluded. Multiple arguments can be 
passed to `--exclude-dir` to exclude multiple subfolders. 

The argument(s) given to `--exclude-dir` may be regular expressions. These regular expressions are 
matched only against the directory name, not the file name.

#### Examples

Exclude subfolder `$IMAGE_ROOT/foo`:
```shell
$ find-dups $IMAGE_ROOT --exclude-dir $IMAGE_ROOT/foo
```
Exclude all subfolders named `foo` or `bar`:
```shell
$ find-dups $IMAGE_ROOT --exclude-dir foo bar
```

### Slow execution

`find-dups` can also use an alternative algorithm which exhaustively compares all images to each
other, being O(N<sup>2</sup>) in the number of images. This algorithm is selected automatically if
`--max-distance` is not 0.

You can use the `--slow` option to use this alternative algorithm specifically. The `--slow` switch
is mutually exclusive with the `--group` switch.

### Progress bar and verbosity control

- `--progress` prints a progress bar each for the process of reading the images, and the process of 
  finding duplicates among the scanned image
- `--debug` prints debugging output
- `--quiet` decreases the log level by 1 for each time it is called; `--debug` and `--quiet` cancel
  each other out

## Development notes

Needs Python3, Pillow imaging library and `pillow-heif` HEIF plugin to run, additionally Wand for 
the test suite.

Uses Poetry for dependency management.

### Installation

From source:
```shell
$ git clone https://gitlab.com/lilacashes/DuplicateImages.git
$ cd DuplicateImages
$ pip3 install poetry
$ poetry install
```

### Running

```shell
$ poetry run find-dups $PICTURE_DIR
```
or
```shell
$ poetry run find-dups -h
```
for a list of all possible options.

### Test suite

Running it all:
```shell
$ poetry run pytest
$ poetry run mypy duplicate_images tests
$ poetry run flake8
$ poetry run pylint duplicate_images tests
$ poetry run bandit -r duplicate_images
```
or simply 
```shell
$ .git_hooks/pre-push
```
Setting the test suite to be run before every push:
```shell
$ cd .git/hooks
$ ln -s ../../.git_hooks/pre-push .
```

### Publishing

A tag is created and the new version is published automatically by GitLab CI on every successful
merge to `master`.

#### Prerequisites

For every Merge Request to `master` it is checked that:
- the `version` number in `pyproject.toml` is not an already existing git tag
- the `CHANGELOG.md` contains an entry for the current version number

#### PyPI

There is a job in GitLab CI for publishing to `pypi.org` that runs as soon as a new tag is added, 
which happens automatically whenever a MR is merged. The tag is the same as the `version` in the 
`pyproject.toml` file. For every MR it needs to be ensured that the `version` is not the same as an 
already existing tag.

To publish the package on PyPI manually:
```shell
$ poetry config repositories.testpypi https://test.pypi.org/legacy/
$ poetry build
$ poetry publish --username $PYPI_USER --password $PYPI_PASSWORD --repository testpypi && \
  poetry publish --username $PYPI_USER --password $PYPI_PASSWORD
```
(obviously assuming here that username and password are the same on PyPI and TestPyPI)

#### Updating GitHub mirror

The GitHub repo `git@github.com:lene/DuplicateImages.git` is set up as a push mirror in GitLab CI, 
but mirroring is flaky at the time and may or may not succeed. The CI job `PushToGithub` should take
care of mirroring to GitHub after every merge to `master`.

To push to the GitHub repository manually (assuming the GitHub repo is set up as remote `github`):
```shell
$ git checkout master
$ git fetch
$ git pull --rebase
$ git tag  # to check that the latest tag is present
$ git push --tags github master 
```

#### Creating Releases on GitHub

The CI job `CreateGithubRelease` creates a Release on GitHub, which can then be found under
https://github.com/lene/DuplicateImages/releases.

### Profiling

#### CPU time
To show the top functions by time spent, including called functions:
```shell
$ poetry run python -m cProfile -s tottime ./duplicate_images/duplicate.py \ 
    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1 | head -n 15
```
or, to show the top functions by time spent in the function alone:
```shell
$ poetry run python -m cProfile -s cumtime ./duplicate_images/duplicate.py \ 
    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1 | head -n 15
```

#### Memory usage
```shell
$ poetry run fil-profile run ./duplicate_images/duplicate.py \
    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1
```
This will open a browser window showing the functions using the most memory (see 
https://pypi.org/project/filprofiler for more details).

## Contributors

- Lene Preuss (https://github.com/lene): primary developer
- Mike Reiche (https://github.com/mreiche): support for arbitrary actions, speedups
- https://github.com/beijingjazzpanda: bug fix

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/lene/DuplicateImages",
    "name": "duplicate_images",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.9",
    "maintainer_email": "",
    "keywords": "",
    "author": "Lene Preuss",
    "author_email": "lene.preuss@gmail.com",
    "download_url": "https://files.pythonhosted.org/packages/09/d5/d08538fda8ebd2b4bae6daccb8968e04a413643052e3dbf364afa3b2bdf3/duplicate_images-0.11.0.tar.gz",
    "platform": null,
    "description": "# Finding Duplicate Images\n\nFinds equal or similar images in a directory containing (many) image files.\n\nOfficial home page: https://github.com/lene/DuplicateImages\n\nDevelopment page: https://gitlab.com/lilacashes/DuplicateImages\n\nPyPI page: https://pypi.org/project/duplicate-images\n\n## Usage\n\nInstalling:\n```shell\n$ pip install duplicate_images\n```\n\nPrinting the help screen:\n```shell\n$ find-dups -h\n```\n\nQuick test run:\n```shell\n$ find-dups $IMAGE_ROOT \n```\n\nTypical usage:\n```shell\n$ find-dups $IMAGE_ROOT --parallel --progress --hash-db hashes.json\n```\n\n### Supported image formats\n\n* JPEG and PNG (tested quite thoroughly)\n* HEIC (experimental support, tested cursorily only)\n* All other \n  [formats supported](https://pillow.readthedocs.io/en/latest/handbook/image-file-formats.html) by \n  the `pillow` Python Imaging Library should work, but are not specifically tested. \n\n#### Explicitly allow huge images\n\nThe `PIL` image library, which is used as backend, limits the size of images to 178956970 pixels by\ndefault, to guard against memory exhaustion. For larger images, specify the maximum image size in \npixels with the `--max-image-pixels` option.\n\n### Image comparison algorithms\n\nUse the `--algorithm` option to select how equal images are found. The default algorithm is `phash`.\n\n`ahash`, `colorhash`, `dhash`, `dhash_vertical`, `phash`, `phash_simple`, `whash`: seven different \nimage hashing algorithms. See https://pypi.org/project/ImageHash for an introduction on image \nhashing and https://tech.okcupid.com/evaluating-perceptual-image-hashes-okcupid for some gory \ndetails which image hashing algorithm performs best in which situation. For a start I recommend \nusing `phash`, and only evaluating the other algorithms if `phash` does not perform satisfactorily \nin your use case.\n\n### Image similarity threshold configuration\n\nUse the `--hash-size` parameter to tune the precision of the hashing algorithms. For the `colorhash`\nalgorithm the hash size is interpreted as the number of bin bits and defaults to 3. For all other\nalgorithms the hash size defaults to 8. For `whash` it must be a power of 2.\n\nUse the `--max-distance` parameter to tune how close images should be to be considered duplicates.\nThe argument is a positive integer. Its value is highly dependent on the algorithm used and the \nnature of the images compared, so the best value for your use case can oly be found through \nexperimentation.\n\n**NOTE:** using the `--max-distance` parameter slows down the comparison considerably with large\nimage collections, making the runtime complexity go from O(N) to O(N<sup>2</sup>). If you want to \nscan collections with at least thousands of images, it is highly recommended to tune the desired \nsimilarity threshold with the `--hash-size` parameter alone, if that is at all possible.\n\n### Pre-storing and using image hashes to speed up computation\n\nUse the `--hash-db ${FILE}.json` or `--hash-db ${FILE}.pickle` option to store image hashes in the \nfile `$FILE` in JSON or Pickle format and read image hashes from that file if they are already \npresent there. This avoids having to compute the image hashes anew at every run and can \nsignificantly speed up run times.\n\n### Handling matching images either as pairs or as groups\n\nBy default, matching images are presented as pairs. With the `--group` CLI option, they are handled\nas a group containing all matching images.\n\nExample: `1.jpg`, `2.jpg` and `3.jpg` in the current folder `.` are equal.\n\n```shell\n$ find-dups .\n1.jpg 2.jpg\n1.jpg 3.jpg\n2.jpg 3.jpg\n$ find-dups . --group\n1.jpg 2.jpg 3.jpg\n```\n\n### Actions for matching image groups\n\nUse the `--on-equal` option to select what to do to pairs of equal images. The default action is \n`print`.\n- `delete-first` or `d1`: deletes the first of the files in the group\n- `delete-last` or `dl`: deletes the last of the files in the group\n- `delete-biggest` or `d>`: deletes the file with the biggest size\n- `delete-smallest` or `d<`: deletes the file with the smallest size\n- `symlink-smaller`: delete the smaller files and replace them to a symlink to the biggest file\n- `eog`: launches the `eog` image viewer to compare the files in the group (*deprecated* by `exec`)\n- `xv`: launches the `xv` image viewer to compare the files in the group (*deprecated* by `exec`)\n- `print`: prints the files in the group\n- `print_inline`: like `print` but without newline\n- `quote`: prints the files in the group quoted for POSIX shells\n- `quote_inline`: like `quote` but without newline\n- `exec`: executes a command (see `--exec` argument below)\n- `none`: does nothing; may be useful for benchmarking and testing\n\nThe `--exec` argument allows calling another program when the `--on-equal exec` option is given.\nYou can pass a command line string like `--exec \"program {1} {2}\"` where `{1}` and `{2}` are\nreplaced by the matching pair files (or first two files in a group), quoted so the shell recognizes\nthe files properly. The wildcard `{*}` expands to all files in a matching group, which when called\nwith the `--group` argument may be more than two images considered equal.\n\n#### Examples:\n* `--exec \"open -a Preview -W {1} {2}\"`: Opens the files in MacOS Preview app and waits for it.\n* `--exec \"ls -s {*}\"`: Prints the size (in blocks) next to all files.\n* `--exec 'for i in {*}; do dirname $i; basename $i; done'`: Shows the directory and the filename\n  separately for all files.\n\n### Parallel execution\n\nUse the `--parallel` option to utilize all free cores on your system for calculating image hashes.\nOptionally, you can specify the number of processes to use with `--parallel $N`.\n\nTo execute the `--on-equal` actions in parallel, use the `--parallel-actions` option, which also can\ntake an optional number of processes to use as argument.\n\n### Excluding subfolders\n\nUse the `--exclude-dir` option to exclude subfolders of `$IMAGE_ROOT` from the search. The argument\nis a regular expression matching the subfolder names to be excluded. Multiple arguments can be \npassed to `--exclude-dir` to exclude multiple subfolders. \n\nThe argument(s) given to `--exclude-dir` may be regular expressions. These regular expressions are \nmatched only against the directory name, not the file name.\n\n#### Examples\n\nExclude subfolder `$IMAGE_ROOT/foo`:\n```shell\n$ find-dups $IMAGE_ROOT --exclude-dir $IMAGE_ROOT/foo\n```\nExclude all subfolders named `foo` or `bar`:\n```shell\n$ find-dups $IMAGE_ROOT --exclude-dir foo bar\n```\n\n### Slow execution\n\n`find-dups` can also use an alternative algorithm which exhaustively compares all images to each\nother, being O(N<sup>2</sup>) in the number of images. This algorithm is selected automatically if\n`--max-distance` is not 0.\n\nYou can use the `--slow` option to use this alternative algorithm specifically. The `--slow` switch\nis mutually exclusive with the `--group` switch.\n\n### Progress bar and verbosity control\n\n- `--progress` prints a progress bar each for the process of reading the images, and the process of \n  finding duplicates among the scanned image\n- `--debug` prints debugging output\n- `--quiet` decreases the log level by 1 for each time it is called; `--debug` and `--quiet` cancel\n  each other out\n\n## Development notes\n\nNeeds Python3, Pillow imaging library and `pillow-heif` HEIF plugin to run, additionally Wand for \nthe test suite.\n\nUses Poetry for dependency management.\n\n### Installation\n\nFrom source:\n```shell\n$ git clone https://gitlab.com/lilacashes/DuplicateImages.git\n$ cd DuplicateImages\n$ pip3 install poetry\n$ poetry install\n```\n\n### Running\n\n```shell\n$ poetry run find-dups $PICTURE_DIR\n```\nor\n```shell\n$ poetry run find-dups -h\n```\nfor a list of all possible options.\n\n### Test suite\n\nRunning it all:\n```shell\n$ poetry run pytest\n$ poetry run mypy duplicate_images tests\n$ poetry run flake8\n$ poetry run pylint duplicate_images tests\n$ poetry run bandit -r duplicate_images\n```\nor simply \n```shell\n$ .git_hooks/pre-push\n```\nSetting the test suite to be run before every push:\n```shell\n$ cd .git/hooks\n$ ln -s ../../.git_hooks/pre-push .\n```\n\n### Publishing\n\nA tag is created and the new version is published automatically by GitLab CI on every successful\nmerge to `master`.\n\n#### Prerequisites\n\nFor every Merge Request to `master` it is checked that:\n- the `version` number in `pyproject.toml` is not an already existing git tag\n- the `CHANGELOG.md` contains an entry for the current version number\n\n#### PyPI\n\nThere is a job in GitLab CI for publishing to `pypi.org` that runs as soon as a new tag is added, \nwhich happens automatically whenever a MR is merged. The tag is the same as the `version` in the \n`pyproject.toml` file. For every MR it needs to be ensured that the `version` is not the same as an \nalready existing tag.\n\nTo publish the package on PyPI manually:\n```shell\n$ poetry config repositories.testpypi https://test.pypi.org/legacy/\n$ poetry build\n$ poetry publish --username $PYPI_USER --password $PYPI_PASSWORD --repository testpypi && \\\n  poetry publish --username $PYPI_USER --password $PYPI_PASSWORD\n```\n(obviously assuming here that username and password are the same on PyPI and TestPyPI)\n\n#### Updating GitHub mirror\n\nThe GitHub repo `git@github.com:lene/DuplicateImages.git` is set up as a push mirror in GitLab CI, \nbut mirroring is flaky at the time and may or may not succeed. The CI job `PushToGithub` should take\ncare of mirroring to GitHub after every merge to `master`.\n\nTo push to the GitHub repository manually (assuming the GitHub repo is set up as remote `github`):\n```shell\n$ git checkout master\n$ git fetch\n$ git pull --rebase\n$ git tag  # to check that the latest tag is present\n$ git push --tags github master \n```\n\n#### Creating Releases on GitHub\n\nThe CI job `CreateGithubRelease` creates a Release on GitHub, which can then be found under\nhttps://github.com/lene/DuplicateImages/releases.\n\n### Profiling\n\n#### CPU time\nTo show the top functions by time spent, including called functions:\n```shell\n$ poetry run python -m cProfile -s tottime ./duplicate_images/duplicate.py \\ \n    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1 | head -n 15\n```\nor, to show the top functions by time spent in the function alone:\n```shell\n$ poetry run python -m cProfile -s cumtime ./duplicate_images/duplicate.py \\ \n    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1 | head -n 15\n```\n\n#### Memory usage\n```shell\n$ poetry run fil-profile run ./duplicate_images/duplicate.py \\\n    --algorithm $ALGORITHM --action-equal none $IMAGE_DIR 2>&1\n```\nThis will open a browser window showing the functions using the most memory (see \nhttps://pypi.org/project/filprofiler for more details).\n\n## Contributors\n\n- Lene Preuss (https://github.com/lene): primary developer\n- Mike Reiche (https://github.com/mreiche): support for arbitrary actions, speedups\n- https://github.com/beijingjazzpanda: bug fix\n",
    "bugtrack_url": null,
    "license": "",
    "summary": "Finds equal or similar images in a directory containing (many) image files",
    "version": "0.11.0",
    "project_urls": {
        "Homepage": "https://github.com/lene/DuplicateImages",
        "Repository": "https://github.com/lene/DuplicateImages.git"
    },
    "split_keywords": [],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "e1a94e867e7d59d1f49178ec812849a92ad60a9ffc80f55c00f8c4a75d81df74",
                "md5": "42185945d2bb634c3826707c35cebede",
                "sha256": "c780f11b9bdf811b997ef0dff2977352f33d449380ee5f490f16dbfa2edb8c89"
            },
            "downloads": -1,
            "filename": "duplicate_images-0.11.0-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "42185945d2bb634c3826707c35cebede",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.9",
            "size": 20067,
            "upload_time": "2024-01-25T21:54:33",
            "upload_time_iso_8601": "2024-01-25T21:54:33.468690Z",
            "url": "https://files.pythonhosted.org/packages/e1/a9/4e867e7d59d1f49178ec812849a92ad60a9ffc80f55c00f8c4a75d81df74/duplicate_images-0.11.0-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "09d5d08538fda8ebd2b4bae6daccb8968e04a413643052e3dbf364afa3b2bdf3",
                "md5": "8844fb014d2bf4ea5d18d4717867c744",
                "sha256": "d9f1f21c365cf9aa4c2c774b581461bec8fc8c7aeb7f0569d5cdf0c0ac0415b2"
            },
            "downloads": -1,
            "filename": "duplicate_images-0.11.0.tar.gz",
            "has_sig": false,
            "md5_digest": "8844fb014d2bf4ea5d18d4717867c744",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.9",
            "size": 18648,
            "upload_time": "2024-01-25T21:54:35",
            "upload_time_iso_8601": "2024-01-25T21:54:35.704152Z",
            "url": "https://files.pythonhosted.org/packages/09/d5/d08538fda8ebd2b4bae6daccb8968e04a413643052e3dbf364afa3b2bdf3/duplicate_images-0.11.0.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-01-25 21:54:35",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "lene",
    "github_project": "DuplicateImages",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "lcname": "duplicate_images"
}
        
Elapsed time: 0.16540s