cloud-volume


Namecloud-volume JSON
Version 8.32.1 PyPI version JSON
download
home_pagehttps://github.com/seung-lab/cloud-volume/
SummaryA serverless client for reading and writing Neuroglancer Precomputed volumes both locally and on cloud services.
upload_time2024-04-10 06:08:46
maintainerNone
docs_urlNone
authorWilliam Silversmith, Nico Kemnitz, Ignacio Tartavull, and others
requires_python~=3.6
licenseLicense :: OSI Approved :: BSD License
keywords neuroglancer volumetric-data numpy connectomics microscopy image-processing biomedical-image-processing s3 gcs mesh meshes skeleton skeletons
VCS
bugtrack_url
requirements boto3 chardet cloud-files compressed-segmentation compresso crackle-codec DracoPy fastremap fpzip gevent google-auth google-cloud-core google-cloud-storage json5 jsonschema numpy networkx python-jsonschema-objects pathos protobuf pyspng-seunglab python-dateutil requests psutil pysimdjson simplejpeg six tenacity tqdm urllib3 zfpc
Travis-CI
coveralls test coverage No coveralls.
            [![Build Status](https://travis-ci.org/seung-lab/cloud-volume.svg?branch=master)](https://travis-ci.org/seung-lab/cloud-volume) [![PyPI version](https://badge.fury.io/py/cloud-volume.svg)](https://badge.fury.io/py/cloud-volume) [![SfN 2018 Poster](https://img.shields.io/badge/poster-SfN%202018-blue.svg)](https://drive.google.com/open?id=1RKtaAGV2f7F13opnkQfbp6YBqmoD3fZi) [![codecov](https://img.shields.io/badge/codecov-link-%23d819a6)](https://codecov.io/gh/seung-lab/cloud-volume) [![DOI](https://zenodo.org/badge/98333149.svg)](https://zenodo.org/badge/latestdoi/98333149)

# CloudVolume: IO for Neuroglancer Datasets

```python
from cloudvolume import CloudVolume

vol = CloudVolume('gs://mylab/mouse/image', parallel=True, progress=True)
image = vol[:,:,:] # Download a whole image stack into a numpy array from the cloud
vol[:,:,:] = image # Upload an entire image stack from a numpy array to the cloud

label = 1
mesh = vol.mesh.get(label)
skel = vol.skeleton.get(label)
```

CloudVolume is a serverless Python client for random access reading and writing of [Neuroglancer](https://github.com/google/neuroglancer/) volumes in "[Precomputed](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed)" format, a set of representations for arbitrarily large volumetric images, meshes, and skeletons. CloudVolume is typically paired with [Igneous](https://github.com/seung-lab/igneous), a Kubernetes compatible system for generating image hierarchies, meshes, skeletons, and other dependency free jobs that can be applied to petavoxel scale images.

Precomputed volumes are typically stored on [AWS S3](https://aws.amazon.com/s3/), [Google Storage](https://cloud.google.com/storage/), or locally. CloudVolume can read and write to these object storage providers given a service account token with appropriate permissions. However, these volumes can be stored on any service, including an ordinary webserver or local filesystem, that supports key-value access.

The combination of [Neuroglancer](https://github.com/google/neuroglancer/), [Igneous](https://github.com/seung-lab/igneous), and CloudVolume comprises a system for visualizing, processing, and sharing (via browser viewable URLs) petascale datasets within and between laboratories. A typical example usage would be to visualize raw electron microscope scans of mouse, fish, or fly brains up to a cubic millimeter in physical dimension. Neuroglancer and Igneous would enable you to visualize each step of the process of montaging the image, fine tuning alignment vector fields, creating segmentation layers, ROI masks, or performing other types of analysis. CloudVolume enables you to read from and write to each of these layers. Recently, we have introduced the ability to interact with the graph server ("PyChunkGraph") that backs proofreading automated segmentations via the `graphene://` format.

You can find a collection of CloudVolume accessible and Neuroglancer viewable datasets at https://neurodata.io/project/ocp/, an open data project by some of our collaborators.

## Highlights

- Random access to petavoxel Neuroglancer images, meshes, and skeletons.
- Nearly all output is immediately visualizable using Neuroglancer.\*
- Reads graph server backed proofreading volumes (via `graphene://`).
- Serverless (except `graphene://`) and multi-cloud.

### Detailed Highlights

- Multi-threaded, supports multi-process and green threads.
- Memory optimized, supports shared memory.
- Lossless connectomics relevant codecs ([`compressed_segmentation`](https://github.com/seung-lab/compressedseg), [`compresso`](https://github.com/seung-lab/compresso), [`crackle`](https://github.com/seung-lab/crackle) (BETA), [`fpzip`](https://github.com/seung-lab/fpzip/), [`zfpc`](https://github.com/seung-lab/zfpc), [`png`](https://en.wikipedia.org/wiki/Portable_Network_Graphics), and [`brotli`](https://en.wikipedia.org/wiki/Brotli))
- Understands image hierarchies & anisotropic pixel resolutions.
- Accomodates downloading missing tiles (`fill_missing=True`).
- Accomodates uploading compressed black tiles to erasure coded file systems (`delete_black_uploads=True`).
- Growing support for the Neuroglancer [sharded format](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed) which dramatically condenses the number of files required to represent petascale datasets, similar to [Cloud Optimized GeoTIFF](https://www.cogeo.org/), which can result in [dramatic cost savings](https://github.com/seung-lab/kimimaro/wiki/The-Economics:-Skeletons-for-the-People).
- Reads Precomputed meshes and skeletons.
- Includes viewers for small images, meshes, and skeletons.
- Only 3 dimensions + RBG channels currently supported for images.
- No data versioning.

## Setup

Cloud-volume is regularly tested on Ubuntu with 3.7, 3.8, 3.9 and 3.10. We officially support Linux and Mac OS. Windows is community supported. After installation, you'll also need to set up your cloud credentials if you're planning on writing files or reading from a private dataset. Once you're finished setting up, you can try [reading from a public dataset](https://github.com/seung-lab/cloud-volume/wiki/Reading-Public-Data-Examples).

#### `pip` Binary Installation

```bash
pip install cloud-volume # standard installation
```

CloudVolume depends on several PyPI packages which are Cython bindings for C++. We have provided compiled binaries for many platforms and python versions, however if you are on an unsupported system, pip will attempt to install from source. In that case, follow the instructions below.

**Windows Note:** If you get errors related to a missing C++ compiler, this blog post might help you: https://www.scivision.dev/python-windows-visual-c-14-required/

#### Optional Dependencies

| Tag             | Description                             | Dependencies          |
|-----------------|-----------------------------------------|-----------------------|
| boss            | `boss://` format support                | intern                |
| test            | Supports testing                        | pytest                |
| mesh_viewer     | `mesh.viewer()` GUI                     | vtk                   |
| skeleton_viewer | `skeleton.viewer()` GUI                 | matplotlib            |
| all_viewers     | All viewers now and in the future.      | vtk, matplotlib       |
| dask            | Supports converting to/from dask arrays | dask\[array\]         |

Example:

```bash
pip install cloud-volume[boss,test,all_viewers]
```

#### `pip` Source Installation

*C++ compiler required.*

```bash
sudo apt-get install g++ python3-dev # python-dev if you're on python2
pip install numpy
pip install cloud-volume
```

Due to packaging problems endemic to Python, Cython packages that depend on numpy require numpy header files be installed before attempting to install the package you want. The numpy headers are not recognized unless numpy is installed in a seperate process that runs first. There are hacks for this issue, but I haven't gotten them to work. If you think binaries should be available for your platform, please let us know by opening an issue.

#### Manual Installation

This can be desirable if you want to hack on CloudVolume itself.

```bash
git clone git@github.com:seung-lab/cloud-volume.git
cd cloud-volume

# With virtualenvwrapper
mkvirtualenv cv
workon cv
# With only virtualenv
virtualenv venv
source venv/bin/activate

sudo apt-get install g++ python3-dev # python-dev if you're on python2
pip install numpy # additional step needed for accelerated compressed_segmentation and fpzip
pip install -e . # without optional dependencies
pip install -e .[all_viewers] # with e.g. the all_viewers optional dependency
```

### Credentials

By default, CloudVolume's configuration and cache files are stored in `$HOME/.cloudvolume` (or in `$HOME/.cloudfiles` since we use CloudFiles for the backend). You can configure where CloudVolume looks for these files with the environment variable `$CLOUD_VOLUME_DIR`.  

Credentials are stored in `$CLOUD_VOLUME_DIR/secrets`. You'll need credentials only for the services you'll use. If you plan to use the local filesystem, you won't need any. For Google Storage ([setup instructions here](https://github.com/seung-lab/cloud-volume/wiki/Setting-up-Google-Cloud-Storage)), default account credentials will be used if available and no service account is provided.

If neither of those two conditions apply, you need a service account credential. If you have your credentials handy, you can provide them like so as a dict, JSON string, or a bare token if the service will accept that.

```python
cv = CloudVolume(..., secrets=...)
```

However, it may be simpler to save your credential to disk so you don't have to always provide it. `google-secret.json` is a service account credential for Google Storage, `aws-secret.json` is a service account for S3, etc. You can support multiple projects at once by prefixing the bucket you are planning to access to the credential filename. `google-secret.json` will be your defaut service account, but if you also want to also access bucket ABC, you can provide `ABC-google-secret.json` and you'll have simultaneous access to your ordinary buckets and ABC. The secondary credentials are accessed on the basis of the bucket name, not the project name.

```bash
mkdir -p ~/.cloudvolume/secrets/
mv aws-secret.json ~/.cloudvolume/secrets/ # needed for Amazon
mv google-secret.json ~/.cloudvolume/secrets/ # needed for Google
mv boss-secret.json ~/.cloudvolume/secrets/ # needed for the BOSS
mv matrix-secret.json ~/.cloudvolume/secrets/ # needed for Matrix
mv tigerdata-secret.json ~/.cloudvolume/secrets/ # needed for Tigerdata
```

#### `aws-secret.json` and `matrix-secret.json`

Create an [IAM user service account](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_users.html) that can read, write, and delete objects from at least one bucket.

```json
{
	"AWS_ACCESS_KEY_ID": "$MY_AWS_ACCESS_KEY_ID",
	"AWS_SECRET_ACCESS_KEY": "$MY_SECRET_ACCESS_TOKEN"
}
```

#### `google-secret.json`

You can create the `google-secret.json` file [here](https://console.cloud.google.com/iam-admin/serviceaccounts). You don't need to manually fill in JSON by hand, the below example is provided to show you what the end result should look like. You should be able to read, write, and delete objects from at least one bucket.

```json
{
  "type": "service_account",
  "project_id": "$YOUR_GOOGLE_PROJECT_ID",
  "private_key_id": "...",
  "private_key": "...",
  "client_email": "...",
  "client_id": "...",
  "auth_uri": "https://accounts.google.com/o/oauth2/auth",
  "token_uri": "https://accounts.google.com/o/oauth2/token",
  "auth_provider_x509_cert_url": "https://www.googleapis.com/oauth2/v1/certs",
  "client_x509_cert_url": ""
}
```

#### `cave-secret.json`

*Note: used to be called chunkedgraph-secret.json. This is still supported but deprecated.*

If you have a token from Graphene/Chunkedgraph server, create the `cave-secret.json` file as shown in the example below. You may also pass the token to `CloudVolume(..., secrets=token)`.

```json
{
  "token": "<your_token>"
}
```

Note that to take advantage of multiple credential files, prepend the fully qualified domain name (FQDN) of the server instead of the bucket for GCS and S3. For example, `sudomain.domain.com-cave-secret.json`.

## Usage

CloudVolume supports reading and writing to Neuroglancer data layers on Amazon S3, Google Storage, The BOSS, and the local file system.

Supported URLs are of the forms:

`$FORMAT://$PROTOCOL://$BUCKET/$DATASET/$LAYER`

The format or protocol fields may be omitted where required. In the case of the precomputed format, the format specifier is optional.

| Format      | Protocols                                    | Default | Example                                |
|-------------|----------------------------------------------|---------|----------------------------------------|
| precomputed | gs, s3, http, https, file, matrix, tigerdata | Yes     | gs://mybucket/dataset/layer            |
| graphene    | gs, s3, http, https, file, matrix, tigerdata |         | graphene://gs://mybucket/dataset/layer |
| boss        | N/A                                          |         | boss://collection/experiment/channel   |
| n5          | gs, s3, http, https, file, matrix, tigerdata |         | n5://gs://mybucket/dataset/layer       |

### Supported Formats

* precomputed: Neuroglancer's native format. ([specification](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed))
* graphene: Precomputed based format used by the PyChunkGraph server.
* boss: The BOSS (https://docs.theboss.io/docs)
* n5: Not HDF5 (https://github.com/saalfeldlab/n5) Read-only support. Supports raw, gzip, bz2, and xz but not lz4 compression. mode 0 datasets only.

### Supported Protocols

* gs:   Google Storage
* s3:   Amazon S3
* http(s): (read-only) Ordinary Web Servers
* file: Local File System (absolute path)
* matrix: Princeton Internal System (run in large part by Seung Lab)
* tigerdata: Princeton Internal System (run by Princeton OIT)

CloudVolume also supports [alternative s3 aliases](https://github.com/seung-lab/cloud-files#alias-for-alternative-s3-endpoints) via CloudFiles.


### `info` Files - New Dataset

Neuroglancer relies on an [`info`](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed#info-json-file-specification) file located at the root of a dataset layer to tell it how to compute file locations and interpret the data in each file. CloudVolume piggy-backs on this functionality.

In the below example, assume you are creating a new segmentation volume from a 3d numpy array "rawdata". Note Precomputed stores data in Fortran (column major, aka CZYX) order. You should do a small test to see if the image is written transposed. You can fix this by uploading `rawdata.T`. A more detailed example for uploading a local volume [is located here](https://github.com/seung-lab/cloud-volume/wiki/Example-Single-Machine-Dataset-Upload).

```python3
from cloudvolume import CloudVolume

info = CloudVolume.create_new_info(
    num_channels    = 1,
    layer_type      = 'segmentation',
    data_type       = 'uint64', # Channel images might be 'uint8'
    # raw, png, jpeg, compressed_segmentation, fpzip, kempressed, zfpc, compresso, crackle
    encoding        = 'raw', 
    resolution      = [4, 4, 40], # Voxel scaling, units are in nanometers
    voxel_offset    = [0, 0, 0], # x,y,z offset in voxels from the origin
    mesh            = 'mesh',
    # Pick a convenient size for your underlying chunk representation
    # Powers of two are recommended, doesn't need to cover image exactly
    chunk_size      = [ 512, 512, 16 ], # units are voxels
    volume_size     = [ 250000, 250000, 25000 ], # e.g. a cubic millimeter dataset
)
vol = CloudVolume(cfg.path, info=info)
vol.commit_info()
vol[cfg.x: cfg.x + cfg.length, cfg.y:cfg.y + cfg.length, cfg.z: cfg.z + cfg.length] = rawdata[:,:,:]
```
| Encoding                | Image Type                 | Lossless | Neuroglancer Viewable | Description                                                                              |
|-------------------------|----------------------------|----------|-------------|------------------------------------------------------------------------------------------|
| raw                     | Any                        | Y        | Y           | Serialized numpy arrays.                                                                 |
| png                     | Image                      | Y        | Y           | Multiple slices stiched into a single PNG.                                               |
| jpeg                    | Image                      | N        | Y           | Multiple slices stiched into a single JPEG.                                              |
| compressed_segmentation | Segmentation               | Y        | Y           | Renumbered numpy arrays to reduce data width. Also used by Neuroglancer internally.      |
| compresso               | Segmentation               | Y        | Y           | Lossless high compression algorithm for connectomics segmentation.                       |
| crackle                 | Segmentation               | Y        | Y*           | Lossless high compression algorithm for connectomics segmentation.                       |
| fpzip                   | Floating Point             | Y        | Y*           | Takes advantage of IEEE 754 structure + L1 Lorenzo predictor to get higher compression.  |
| kempressed              | Anisotropic Z Floating Point | N**      | Y*           | Adds manipulations on top of fpzip to achieve higher compression.                        |
| zfpc                    | Alignment Vector Fields    | N***     | Y*          | zfp stream container.                        |

\* Not integrated into official Neuroglancer yet, but available on a [fork](https://allcodecs-dot-neuromancer-seung-import.appspot.com/) which can be seen on Github [here](https://github.com/william-silversmith/neuroglancer/tree/wms_combined_codecs).
\*\* Lossless if your data can handle adding and then subtracting 2.
\*\*\* Lossless by default, but you probably want to use the lossy mode.

Note on `compressed_segmentation`: To use, make sure `compressed_segmentation_block_size` is specified (usually `[8,8,8]`. This field will appear in the `info` file in the relevant scale.

Note on `zfpc`: To configure, use the fields `zfpc_rate`, `zfpc_precision`, `zfpc_tolerance`, `zfpc_correlated_dims` in the relevant scale of the `info` file.


### Examples

```python
# Basic Examples
vol = CloudVolume('gs://mybucket/retina/image')
vol = CloudVolume('gs://mybucket/retina/image', secrets=token, dict or json)
vol = CloudVolume('gs://bucket/dataset/channel', mip=0, bounded=True, fill_missing=False)
vol = CloudVolume('gs://bucket/dataset/channel', mip=[ 8, 8, 40 ], bounded=True, fill_missing=False) # set mip at this resolution
vol = CloudVolume('gs://bucket/datasset/channel', info=info) # New info file from scratch
image = vol[:,:,:] # Download the entire image stack into a numpy array
image = vol.download(bbox, mip=2, renumber=True) # download w/ smaller dtype
image = vol.download(bbox, mip=2, label=777) # download binary image for label
uniq = vol.unique(bbox, mip=0) # efficient extraction of unique labels
listing = vol.exists( np.s_[0:64, 0:128, 0:64] ) # get a report on which chunks actually exist
exists = vol.image.has_data(mip=0) # boolean check to see if any data is there
listing = vol.delete( np.s_[0:64, 0:128, 0:64] ) # delete this region (bbox must be chunk aligned)
vol[64:128, 64:128, 64:128] = image # Write a 64^3 image to the volume
img = vol.download_point( (x,y,z), size=256, mip=3 ) # download region around (mip 0) x,y,z at mip 3
pts = vol.scattered_points([ (x1,y1,z1), (x2,y2,z2) ]) # download voxel labels located at indicated points
# download image files without decompressing or rendering them. Good for caching!
files = vol.download_files(bbox, mip, decompress=False)

# creates an anonymous in-memory CloudVolume that 
# will self-clean when the reference count drops to zero. 
# Store compressed images in memory for quick access!
mem_vol = vol.image.memory_cutout(bbox, mip=1, encoding="compresso")

# Server
vol.viewer() # launches neuroglancer compatible web server on http://localhost:1337

# Microviewer (outdated, see https://github.com/seung-lab/microviewer/)
img = vol[64:1028, 64:1028, 64:128]
img.viewer() # launches web viewer on http://localhost:8080

# Meshes
vol.mesh.save(12345) # save 12345 as ./12345.ply on disk
vol.mesh.save([12345, 12346, 12347]) # merge three segments into one file
vol.mesh.save(12345, file_format='obj') # 'ply' and 'obj' are both supported
vol.mesh.get(12345) # return the mesh as vertices and faces instead of writing to disk
vol.mesh.get([ 12345, 12346 ]) # return these two segids fused into a single mesh
vol.mesh.get([ 12345, 12346 ], fuse=False) # return { 12345: mesh, 12346: mesh }
vol.mesh.put(meshes) # works for unsharded legacy only
vol.mesh.delete(segids) # works for unsharded meshes only

mesh.viewer() # Opens GUI. Requires vtk.

# Skeletons
skel = vol.skeleton.get(12345)
vol.skeleton.upload_raw(segid, skel.vertices, skel.edges, skel.radii, skel.vertex_types)
vol.skeleton.upload(skel)

# specified in nm, only available for datasets with a generated index
skels = vol.skeleton.get_by_bbox( Bbox( (0,0,0), (500, 500, 500) ) )
vol.skeleton.spatial_index # None if not available

skel.empty() # boolean

bytes = skel.encode() # encode to Precomputed format (bytes)
skel = Skeleton.decode(bytes) # decode from PrecomputedFormat

skel = skel.crop(slices or bbox) # eliminate vertices and edges outside bbox
skel = skel.consolidate() # eliminate duplicate vertices and edges
skel3 = skel.merge(skel2) # merge two skeletons into one
skel = skel.clone() # create copy
skel = Skeleton.from_swc(swcstr) # decode an SWC file
skel_str = skel.to_swc() # convert to SWC file in string representation
skel.viewer() # Opens GUI. Requires matplotlib

skel.cable_length() # sum of all edge lengths
skel = skel.downsample(2) # reduce size of skeleton by factor of 2

skel1 == skel2 # check if contents of internal arrays match
Skeleton.equivalent(skel1, skel2) # ...even if there are differences like differently numbered edges

# Parallel Operation
vol = CloudVolume('gs://mybucket/retina/image', parallel=True) # Use all cores
vol.parallel = 4 # e.g. any number > 1, use this many cores
data = vol[:] # uses shared memory to coordinate processes under the hood

# Shared Memory Output (can be used by other processes)
vol = CloudVolume(...)
# data backed by a shared memory buffer
# location is optional (defaults to vol.shared_memory_id)
data = vol.download_to_shared_memory(np.s_[:], location='some-example')
vol.unlink_shared_memory() # delete the shared memory associated with this cloudvolume
vol.shared_memory_id # get/set the default shared memory location for this instance

# Shared Memory Upload
vol = CloudVolume(...)
vol.upload_from_shared_memory('my-shared-memory-id', # do not prefix with /dev/shm
    bbox=Bbox( (0,0,0), (10000, 7500, 64) ))

# Download or Upload directly with Files
# The files must be in Precomputed raw format.
vol.download_to_file('/path/to/file', bbox=Bbox(...), mip=0) # bbox is the download region
vol.upload_from_file('/path/to/file', bbox=Bbox(...), mip=0) # bbox is the region it represents

# Transfer w/o Excess Memory Allocation
vol = CloudVolume(...)
# single core, send all of vol to destination, no painting memory
# you can also transcode the image encoding and compression type
vol.transfer_to('gs://bucket/dataset/layer', vol.bounds)

# Caching, default located at $HOME/.cloudvolume/cache/$PROTOCOL/$BUCKET/$DATASET/$LAYER/$RESOLUTION
# You can also set the cache location using
# cache=str or with environment variable CLOUD_VOLUME_CACHE_DIR
vol = CloudVolume('gs://mybucket/retina/image', cache=True) # Basic Example
image = vol[0:10,0:10,0:10] # Download partial image and cache
vol[0:10,0:10,0:10] = image # Upload partial image and cache

# Resizing and clearing the LRU in-memory cache
vol = CloudVolume(..., lru_bytes=num_bytes) # >= 0, 0 means disabled
vol.image.lru.resize(num_bytes) # same
vol.image.lru.clear()
len(vol.image.lru) # number of items in lru
vol.image.lru.nbytes # size in bytes (not counting LRU structures, nor recursive)
vol.image.lru.items() # etc, also functions as a dict

# Evaluating the on-disk Cache
vol.cache.list() # list files in cache at this mip level
vol.cache.list(mip=1) # list files in cache at mip 1
vol.cache.list_meshes()
vol.cache.list_skeletons()
vol.cache.num_files() # number of files at this mip level
vol.cache.num_bytes(all_mips=True) # Return num files for each mip level in a list
vol.cache.num_bytes() # number of bytes taken up by files, size on disk can be bigger
vol.cache.num_bytes(all_mips=True) # Return num bytes for each mip level in a list

vol.cache.enabled = True/False # Turn the cache on/off
vol.cache.path = Str # set the cache location
vol.cache.compress = None/True/False # None: Link to cloud setting, Boolean: Force cache to compressed (True) or uncompressed (False)

# Deleting Cache
vol.cache.flush() # Delete local cache for this layer at this mip level
vol.cache.flush(preserve=Bbox(...)) # Same, but preserve cache in a region of space
vol.cache.flush_region(region=Bbox(...), mips=[...]) # Delete the cached files in this region at these mip levels (default all mips)
vol.cache.flush_info()
vol.cache.flush_provenance()

# Using Green Threads
import gevent.monkey
gevent.monkey.patch_all(thread=False)

cv = CloudVolume(..., green_threads=True)
img = cv[...] # now green threads will be used

# Dask Interface (requires dask installation)
arr = cv.to_dask()
arr = cloudvolume.dask.from_cloudvolume(cloudpath) # same as to_dask
res = cloudvolume.dask.to_cloudvolume(arr, cloudpath, compute=bool, return_store=bool)
```

### CloudVolume Constructor

```python3
CloudVolume(
    cloudpath:str, mip:int=0, bounded:bool=True, 
    autocrop:bool=False, fill_missing:bool=False, cache:CacheType=False, 
    compress_cache:CompressType=None, cdn_cache:bool=True, 
    progress:bool=INTERACTIVE, info:dict=None, provenance:dict=None,
    compress:CompressType=None, compress_level:Optional[int]=None, 
    non_aligned_writes:bool=False, parallel:ParallelType=1, delete_black_uploads:bool=False, 
    background_color:int=0, green_threads:bool=False, use_https:bool=False,
    max_redirects:int=10, mesh_dir:Optional[str]=None, skel_dir:Optional[str]=None, 
    agglomerate:bool=False, secrets:SecretsType=None, 
    spatial_index_db:Optional[str]=None, lru_bytes:int = 0,
    cache_locking:bool = True
)
```


*      agglomerate: (bool, graphene only) sets the default mode for downloading
        images to agglomerated (True) vs watershed (False).
*      autocrop: (bool) If the specified retrieval bounding box exceeds the
          volume bounds, process only the area contained inside the volume. 
          This can be useful way to ensure that you are staying inside the 
          bounds when `bounded=False`.
*      background_color: (number) Specifies what the "background value" of the
        volume is (traditionally 0). This is mainly for changing the behavior
        of delete_black_uploads.
*      bounded: (bool) If a region outside of volume bounds is accessed:
          True: Throw an error
          False: Allow accessing the region. If no files are present, an error 
              will still be thrown. Consider combining this option with 
              `fill_missing=True`. However, this can be dangrous as it allows
              missing files and potentially network errors to be intepreted as 
              zeros.
*      cache: (bool or str) Store downs and uploads in a cache on disk
            and preferentially read from it before redownloading.
          - falsey value: no caching will occur.
          - True: cache will be located in a standard location.
          - non-empty string: cache is located at this file path

          After initialization, you can adjust this setting via:
          `cv.cache.enabled = ...` which accepts the same values.

          Note: This cache is totally separate from the LRU controlled by 
          lru_bytes.
*      cache_locking: (bool) The local cache will use file locks via fasteners to prevent issues with multi-process cache access. If this is not a concern, performance can be slightly improved by setting this to False. This uses CloudFiles' locking mechanism.
*      cdn_cache: (int, bool, or str) Sets Cache-Control HTTP header on uploaded 
        image files. Most cloud providers perform some kind of caching. As of 
        this writing, Google defaults to 3600 seconds. Most of the time you'll 
        want to go with the default. 
        - int: number of seconds for cache to be considered fresh (max-age)
        - bool: True: max-age=3600, False: no-cache
        - str: set the header manually
*      compress: (bool, str, None) pick which compression method to use.
*          None: (default) gzip for raw arrays and no additional compression
            for compressed_segmentation and fpzip.
          bool: 
            True=gzip, 
            False=no compression, Overrides defaults
          str: 
            'gzip': Extension so that we can add additional methods in the future 
                    like lz4 or zstd. 
            'br': Brotli compression, better compression rate than gzip
            '': no compression (same as False).
*      compress_level: (int, None) level for compression. Higher number results
          in better compression but takes longer.
        Defaults to 9 for gzip (ranges from 0 to 9).
        Defaults to 5 for brotli (ranges from 0 to 11).
*      compress_cache: (None or bool) If not None, override default compression 
          behavior for the cache.
*      delete_black_uploads: (bool) If True, on uploading an entirely black chunk,
          issue a DELETE request instead of a PUT. This can be useful for avoiding storing
          tiny files in the region around an ROI. Some storage systems using erasure coding 
          don't do well with tiny file sizes.
*      fill_missing: (bool) If a chunk file is unable to be fetched:
          True: Use a block of zeros
          False: Throw an error
*      green_threads: (bool) Use green threads instead of preemptive threads. This
        can result in higher download performance for some compression types. Preemptive
        threads seem to reduce performance on multi-core machines that aren't densely
        loaded as the CPython threads are assigned to multiple cores and the thrashing
        + GIL reduces performance. You'll need to add the following code to the top
        of your program to use green threads:

            import gevent.monkey
            gevent.monkey.patch_all(threads=False)
*      lru_bytes: (int) number of bytes used to cache recently used image 
        tiles in memory. This is an in-memory cache and is completely separate from
        the `cache` parameter that handles disk IO. Tiles are stripped over only their
        second stage compression.
*      info: (dict) In lieu of fetching a neuroglancer info file, use this one.
          This is useful when creating new datasets and for repeatedly initializing
          a new cloudvolume instance.
*      max_redirects: (int) if > 0, allow up to this many redirects via info file 'redirect'
          data fields. If <= 0, allow no redirections and access the current info file directly
          without raising an error.
*      mesh_dir: (str) if not None, override the info['mesh'] key before pulling the
        mesh info file.
*      mip: (int or iterable) Which level of downsampling to read and write from.
          0 is the highest resolution. You can also specify the voxel resolution
          like mip=[6,6,30] which will search for the appropriate mip level.
*      non_aligned_writes: (bool) Enable non-aligned writes. Not multiprocessing 
          safe without careful design. When not enabled, a 
          cloudvolume.exceptions.AlignmentError is thrown for non-aligned writes. 
          
          https://github.com/seung-lab/cloud-volume/wiki/Advanced-Topic:-Non-Aligned-Writes

      parallel (int: 1, bool): Number of extra processes to launch, 1 means only 
          use the main process. If parallel is True use the number of CPUs 
          returned by multiprocessing.cpu_count(). When parallel > 1, shared
          memory (Linux) or emulated shared memory via files (other platforms) 
          is used by the underlying download.
*      progress: (bool) Show progress bars. 
          Defaults to True in interactive python, False in script execution mode.
*      provenance: (string, dict) In lieu of fetching a provenance 
          file, use this one. 
*      secrets: (dict) provide per-instance authorization tokens. If not provided,
        defaults to looking in .cloudvolume/secrets for necessary tokens.
*      skel_dir: (str) if not None, override the info['skeletons'] key before 
        pulling the skeleton info file.
*      spatial_index_db: (str) A path to an sqlite3 or mysql database that follows 
        the following uri schema. sqlite is assumed if no scheme is present in 
        the uri.
          [sqlite://]filename.db
          mysql://<username>:<password>@<host>:<port>/<db_name>

        Igneous generated datasets include a JSON based spatial
        database that tiles the dataset. This can be fast enough up to about 100 TVx
        datasets. Above that, a proper database is required for efficient queries.
        We provide multiple SQL database types that the index can be hosted on.
*      use_https: (bool) maps gs:// and s3:// to their respective https paths. The 
        https paths hit a cached, read-only version of the data and may be faster.

### CloudVolume Methods

Better documentation coming later, but for now, here's a summary of the most useful method calls. Use help(cloudvolume.CloudVolume.$method) for more info.

* create_new_info (class method) - Helper function for creating info files for creating new data layers.
* refresh_info - Repull the info file.
* refresh_provenance - Repull the provenance file.
* bbox_to_mip - Covert a bounding box or slice from one mip level to another.
* slices_from_global_coords - *deprecated, why not use bbox_to_mip?* Find the CloudVolume slice from MIP 0 coordinates if you're on a different MIP. Often used in combination with neuroglancer.
* reset_scales - Delete mips other than 0 in the info file. Does not autocommit.
* add_scale - Generate a new mip level in the info property. Does not autocommit.
* commit_info - Push the current info property into the cloud as a JSON file.
* commit_provenance - Push the current provenance property into the cloud as a JSON file.
* image - Access image operations directly.
  * download - Download bounding boxes from a given mip level.
  * upload - Upload images to bounding boxes at a given mip level.
  * transfer_to - Transfer data without painting a container array to avoid out of memory errors.
  * exists - Check which chunk files exist in a given bounding box.
  * delete - Delete chunks in a given bounding box at a given mip level.
* mesh - Access mesh operations
	* get - Download an object. Can merge multiple segmentids
	* save - Download an object and save it in `.obj` format. You can combine equivialences into a single object too.
* skeleton - Access Skeletons
  * get - Download an object.
  * upload - Save a skeleton object to the cloud.
* cache - Access cache operations
	* enabled - Boolean switch to enable/disable cache. If true, on reading, check local disk cache before downloading, and save downloaded chunks to cache. When writing, write to the cloud then save the chunks you wrote to cache. If false, bypass cache completely. The cache is located at `$HOME/.cloudvolume/cache`.
	* path - Property that shows the current filesystem path to the cache
	* list - List files in cache
	* num_files - Number of files in cache at this mip level , use all_mips=True to get them all
	* num_bytes - Return the number of bytes in cache at this mip level, all_mips=True to get them all
	* flush - Delete the cache at this mip level, preserve=Bbox/slice to save a spatial region
	* flush_region - Delete a spatial region at this mip level
* exists - Generate a report on which chunks within a bounding box exist.
* delete - Delete the chunks within this bounding box.
* transfer_to - Transfer data from a bounding box to another data storage location. Does not allocate memory and transfers in blocks, so can transfer large volumes of data. May be less efficient than a dedicated tool like `gsutil` or `aws s3`.
* unlink_shared_memory - Delete shared memory associated with this instance (`vol.shared_memory_id`)
* generate_shared_memory_location - Create a new unique shared memory identifier string. No side effects.
* download_to_shared_memory - Instead of using ordinary numpy memory allocations, download to shared memory.
    Be careful, shared memory is like a file and doesn't disappear unless explicitly unlinked. (`vol.unlink_shared_memory()`)
* upload_from_shared_memory - Upload from a given shared memory block without making a copy.
* download_point - Download the region around this mip 0 coordinate at a given mip level.

### CloudVolume Properties

Accessed as `vol.$PROPERTY` like `vol.mip`. Parens next to each property mean (data type:default, writability). (r) means read only, (w) means write only, (rw) means read/write.

* mip (uint:0, rw) - Read from and write to this mip level (0 is highest res). Each additional increment in the number is typically a 2x reduction in resolution.
* bounded (bool:True, rw) - If a region outside of volume bounds is accessed throw an error if True or Fill the region with black (useful for e.g. marching cubes's 1px boundary) if False.
* autocrop (bool:False, rw) - If bounded is False and this option is True, automatically crop requested uploads and downloads to the volume boundary.
* fill_missing (bool:False, rw) - If a file inside volume bounds is unable to be fetched use a block of zeros if True, else throw an error.
* delete_black_uploads (bool:False, rw) - If True, issue a DELETE http request instead of a PUT when an individual uploaded chunk is all zeros.
* info (dict, rw) - Python dict representation of Neuroglancer info JSON file. You must call `vol.commit_info()` to save your changes to storage.
* provenance (dict-like, rw) - Data layer provenance file representation. You must call `vol.commit_provenance()` to save your changes to storage.
* available_mips (list of ints, r) - Query which mip levels are defined for reading and writing.
* dataset_name (str, rw) - Which dataset (e.g. test_v0, snemi3d_v0) on S3, GS, or FS you're reading and writing to. Known as an "experiment" in BOSS terminology. Writing to this property triggers an info refresh.
* layer (str, rw) - Which data layer (e.g. image, segmentation) on S3, GS, or FS you're reading and writing to. Known as a "channel" in BOSS terminology. Writing to this property triggers an info refresh.
* base_cloudpath (str, r) - The cloud path to the dataset e.g. s3://bucket/dataset/
* layer_cloudpath (str, r) - The cloud path to the data layer e.g. gs://bucket/dataset/image
* info_cloudpath (str, r) - Generate the cloud path to this data layer's info file.
* scales (dict, r) - Shortcut to the 'scales' property of the info object
* scale (dict, rw)* - Shortcut to the working scale of the current mip level
* shape (Vec4, r)* - Like numpy.ndarray.shape for the entire data layer.
* volume_size (Vec3, r)* - Like shape, but omits channel (x,y,z only).
* num_channels (int, r) - The number of channels, the last element of shape.
* layer_type (str, r) - The neuroglancer info type, 'image' or 'segmentation'.
* dtype (str, r) - The info data_type of the volume, e.g. uint8, uint32, etc. Similar to numpy.ndarray.dtype.
* encoding (str, r) - The neuroglancer info encoding. e.g. 'raw', 'jpeg', 'npz'
* resolution (Vec3, r)* - The 3D physical resolution of a voxel in nanometers at the working mip level.
* downsample_ratio (Vec3, r) - Ratio of the current resolution to the highest resolution mip available.
* chunk_size (Vec3, r)* - Size of the underlying chunks that constitute the volume in storage. e.g. Vec(64, 64, 64)
* key (str, r)* - The 'directory' we're accessing the current working mip level from within the data layer. e.g. '6_6_30'
* bounds (Bbox, r)* - A Bbox object that represents the bounds of the entire volume.
* shared_memory_id (str, rw) - Shared memory location used for parallel operation or for output.

\* These properties can also be accessed with a function named like `vol.mip_$PROPERTY($MIP)`. By default they return the current mip level assigned to the CloudVolume, but any mip level can be accessed via the corresponding `mip_` function. Example: `vol.mip_resolution(2)` would return the resolution of mip 2.

### VolumeCutout Functions

When you download an image using CloudVolume it gives you a `VolumeCutout`. These are `numpy.ndarray` subclasses that support a few extra properties to help make book keeping easier. The major advantage is `save_images()` which can help you view your dataset as PNG slices.

* `dataset_name` - The dataset this image came from.
* `layer` - Which layer it came from.
* `mip` - Which mip it came from
* `layer_type` - "image" or "segmentation"
* `bounds` - The bounding box of the cutout
* `num_channels` - Alias for `vol.shape[3]`
* `save_images()` - Save Z slice PNGs of the current image to `./saved_images` for manual inspection
* `viewer()` - Start a local web server (http://localhost:8080) that can view small volumes interactively. This was recently changed from `view` as `view` is a useful numpy method.

### Viewing a Precomputed Volume on Disk

If you have Precomputed volume onto local disk and would like to point neuroglancer to it:

```python
vol = CloudVolume(...)
vol.viewer()
```

You can then point any version of neuroglancer at it using `precomputed://http://localhost:1337/NAME_OF_LAYER`.

### Microviewer

CloudVolume includes a built-in dependency free viewer for 3D volumetric datasets smaller than about 2GB uncompressed. It supports bool, uint8, uint16, uint32, float32, and float64 numpy data types for both images and segmentation and can render a composite overlay of image and segmentation.

You can launch a viewer using the `.viewer()` method of a VolumeCutout object or by using the `view(...)` or `hyperview(...)` functions that come with the cloudvolume module. This launches a web server on `http://localhost:8080`. You can read more [on the wiki](https://github.com/seung-lab/cloud-volume/wiki/%CE%BCViewer).

```python3
from cloudvolume import CloudVolume, view, hyperview

channel_vol = CloudVolume(...)
seg_vol = CloudVolume(...)
img = vol[...]
seg = vol[...]

img.viewer() # works on VolumeCutouts
seg.viewer() # segmentation type derived from info
view(img) # alternative for arbitrary numpy arrays
view(seg, segmentation=True)
hyperview(img, seg) # img and seg shape must match

>>> Viewer server listening to http://localhost:8080
```

There are also seperate viewers for skeleton and mesh objects that can be invoked by calling `.viewer()` on either object. However, skeletons depend on `matplotlib` and meshes depend on `vtk` and OpenGL to function.

```bash
pip install vtk matplotlib
```

## Python 2.7 End of Life

Python 2.7 is no longer supported by CloudVolume. Updated versions of `pip` will download the last supported release 1.21.1. You can read more on the policy page: https://github.com/seung-lab/cloud-volume/wiki/Policy#python-27-end-of-life

## Related Projects

1. [Igneous](https://github.com/seung-lab/igneous): Computational pipeline for visualizing neuroglancer volumes.
2. [CloudVolume.jl](https://github.com/seung-lab/CloudVolume.jl): CloudVolume in Julia
3. [fpzip](https://github.com/seung-lab/fpzip): A Python Package for the C++ code by Lindstrom et al.
4. [compressed_segmentation](https://github.com/seung-lab/compressedseg): A Python Package wrapping the code for the compressed_segmentation format developed by Jeremy Maitin-Shepard and Stephen Plaza.
5. [Kimimaro](https://github.com/seung-lab/kimimaro): High performance skeletonization of densely labeled 3D volumes.
6. [compresso](https://github.com/seung-lab/compresso): High lossless compression of connectomics segmentation. Algorithm by and code derived from Matejek et al.
7. [zfpc](https://github.com/seung-lab/zfpc): Optimized zfp multi-stream container for alignment vector fields (and similar floating point data).
8. [crackle](https://github.com/seung-lab/crackle): Lossless high compression of connectomics segmentation. (BETA)

## Acknowledgments

Thank you to everyone that has contributed past or current to CloudVolume or the ecosystem it serves. We love you!  

Jeremy Maitin-Shepard created [Neuroglancer](https://github.com/google/neuroglancer) and defined the Precomputed format. Yann Leprince provided a [pure Python codec](https://github.com/HumanBrainProject/neuroglancer-scripts) for the compressed_segmentation format. Jeremy Maitin-Shepard and Stephen Plaza created C++ code defining the compression and decompression (respectively) protocol for [compressed_segmentation](https://github.com/janelia-flyem/compressedseg). Peter Lindstrom et al. created [the fpzip algorithm](https://computation.llnl.gov/projects/floating-point-compression), and contributed a C++ implementation and advice. Nico Kemnitz adapted our data to fpzip using the "Kempression" protocol (we named it, not him). Dan Bumbarger contributed code and information helpful for getting CloudVolume working on Windows. Fredrik Kihlander's [pure python implementation](https://github.com/wc-duck/pymmh3) of murmurhash3 and [Austin Appleby](https://github.com/aappleby/smhasher) developed murmurhash3 which is necessary for the sharded format. Ben Falk advocated for and did the bulk of the work on brotli compression. Some of the ideas in CloudVolume are based on work by Jingpeng Wu in [BigArrays.jl](https://github.com/seung-lab/BigArrays.jl).  Sven Dorkenwald, Manuel Castro, and Akhilesh Halageri contributed advice and code towards implementing the graphene interface. Oluwaseun Ogedengbe contributed documentation for the sharded format. Eric Perlman wrote the reader for Neuroglancer Multi-LOD meshes. Ignacio Tartavull and William Silversmith wrote the initial version of CloudVolume.

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/seung-lab/cloud-volume/",
    "name": "cloud-volume",
    "maintainer": null,
    "docs_url": null,
    "requires_python": "~=3.6",
    "maintainer_email": null,
    "keywords": "neuroglancer volumetric-data numpy connectomics microscopy image-processing biomedical-image-processing s3 gcs mesh meshes skeleton skeletons",
    "author": "William Silversmith, Nico Kemnitz, Ignacio Tartavull, and others",
    "author_email": "ws9@princeton.edu",
    "download_url": "https://files.pythonhosted.org/packages/8e/a7/b59826cf80ff1f4507ccee39571aec004ae78a07d62680efdfe288a3987d/cloud-volume-8.32.1.tar.gz",
    "platform": null,
    "description": "[![Build Status](https://travis-ci.org/seung-lab/cloud-volume.svg?branch=master)](https://travis-ci.org/seung-lab/cloud-volume) [![PyPI version](https://badge.fury.io/py/cloud-volume.svg)](https://badge.fury.io/py/cloud-volume) [![SfN 2018 Poster](https://img.shields.io/badge/poster-SfN%202018-blue.svg)](https://drive.google.com/open?id=1RKtaAGV2f7F13opnkQfbp6YBqmoD3fZi) [![codecov](https://img.shields.io/badge/codecov-link-%23d819a6)](https://codecov.io/gh/seung-lab/cloud-volume) [![DOI](https://zenodo.org/badge/98333149.svg)](https://zenodo.org/badge/latestdoi/98333149)\n\n# CloudVolume: IO for Neuroglancer Datasets\n\n```python\nfrom cloudvolume import CloudVolume\n\nvol = CloudVolume('gs://mylab/mouse/image', parallel=True, progress=True)\nimage = vol[:,:,:] # Download a whole image stack into a numpy array from the cloud\nvol[:,:,:] = image # Upload an entire image stack from a numpy array to the cloud\n\nlabel = 1\nmesh = vol.mesh.get(label)\nskel = vol.skeleton.get(label)\n```\n\nCloudVolume is a serverless Python client for random access reading and writing of [Neuroglancer](https://github.com/google/neuroglancer/) volumes in \"[Precomputed](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed)\" format, a set of representations for arbitrarily large volumetric images, meshes, and skeletons. CloudVolume is typically paired with [Igneous](https://github.com/seung-lab/igneous), a Kubernetes compatible system for generating image hierarchies, meshes, skeletons, and other dependency free jobs that can be applied to petavoxel scale images.\n\nPrecomputed volumes are typically stored on [AWS S3](https://aws.amazon.com/s3/), [Google Storage](https://cloud.google.com/storage/), or locally. CloudVolume can read and write to these object storage providers given a service account token with appropriate permissions. However, these volumes can be stored on any service, including an ordinary webserver or local filesystem, that supports key-value access.\n\nThe combination of [Neuroglancer](https://github.com/google/neuroglancer/), [Igneous](https://github.com/seung-lab/igneous), and CloudVolume comprises a system for visualizing, processing, and sharing (via browser viewable URLs) petascale datasets within and between laboratories. A typical example usage would be to visualize raw electron microscope scans of mouse, fish, or fly brains up to a cubic millimeter in physical dimension. Neuroglancer and Igneous would enable you to visualize each step of the process of montaging the image, fine tuning alignment vector fields, creating segmentation layers, ROI masks, or performing other types of analysis. CloudVolume enables you to read from and write to each of these layers. Recently, we have introduced the ability to interact with the graph server (\"PyChunkGraph\") that backs proofreading automated segmentations via the `graphene://` format.\n\nYou can find a collection of CloudVolume accessible and Neuroglancer viewable datasets at https://neurodata.io/project/ocp/, an open data project by some of our collaborators.\n\n## Highlights\n\n- Random access to petavoxel Neuroglancer images, meshes, and skeletons.\n- Nearly all output is immediately visualizable using Neuroglancer.\\*\n- Reads graph server backed proofreading volumes (via `graphene://`).\n- Serverless (except `graphene://`) and multi-cloud.\n\n### Detailed Highlights\n\n- Multi-threaded, supports multi-process and green threads.\n- Memory optimized, supports shared memory.\n- Lossless connectomics relevant codecs ([`compressed_segmentation`](https://github.com/seung-lab/compressedseg), [`compresso`](https://github.com/seung-lab/compresso), [`crackle`](https://github.com/seung-lab/crackle) (BETA), [`fpzip`](https://github.com/seung-lab/fpzip/), [`zfpc`](https://github.com/seung-lab/zfpc), [`png`](https://en.wikipedia.org/wiki/Portable_Network_Graphics), and [`brotli`](https://en.wikipedia.org/wiki/Brotli))\n- Understands image hierarchies & anisotropic pixel resolutions.\n- Accomodates downloading missing tiles (`fill_missing=True`).\n- Accomodates uploading compressed black tiles to erasure coded file systems (`delete_black_uploads=True`).\n- Growing support for the Neuroglancer [sharded format](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed) which dramatically condenses the number of files required to represent petascale datasets, similar to [Cloud Optimized GeoTIFF](https://www.cogeo.org/), which can result in [dramatic cost savings](https://github.com/seung-lab/kimimaro/wiki/The-Economics:-Skeletons-for-the-People).\n- Reads Precomputed meshes and skeletons.\n- Includes viewers for small images, meshes, and skeletons.\n- Only 3 dimensions + RBG channels currently supported for images.\n- No data versioning.\n\n## Setup\n\nCloud-volume is regularly tested on Ubuntu with 3.7, 3.8, 3.9 and 3.10. We officially support Linux and Mac OS. Windows is community supported. After installation, you'll also need to set up your cloud credentials if you're planning on writing files or reading from a private dataset. Once you're finished setting up, you can try [reading from a public dataset](https://github.com/seung-lab/cloud-volume/wiki/Reading-Public-Data-Examples).\n\n#### `pip` Binary Installation\n\n```bash\npip install cloud-volume # standard installation\n```\n\nCloudVolume depends on several PyPI packages which are Cython bindings for C++. We have provided compiled binaries for many platforms and python versions, however if you are on an unsupported system, pip will attempt to install from source. In that case, follow the instructions below.\n\n**Windows Note:** If you get errors related to a missing C++ compiler, this blog post might help you: https://www.scivision.dev/python-windows-visual-c-14-required/\n\n#### Optional Dependencies\n\n| Tag             | Description                             | Dependencies          |\n|-----------------|-----------------------------------------|-----------------------|\n| boss            | `boss://` format support                | intern                |\n| test            | Supports testing                        | pytest                |\n| mesh_viewer     | `mesh.viewer()` GUI                     | vtk                   |\n| skeleton_viewer | `skeleton.viewer()` GUI                 | matplotlib            |\n| all_viewers     | All viewers now and in the future.      | vtk, matplotlib       |\n| dask            | Supports converting to/from dask arrays | dask\\[array\\]         |\n\nExample:\n\n```bash\npip install cloud-volume[boss,test,all_viewers]\n```\n\n#### `pip` Source Installation\n\n*C++ compiler required.*\n\n```bash\nsudo apt-get install g++ python3-dev # python-dev if you're on python2\npip install numpy\npip install cloud-volume\n```\n\nDue to packaging problems endemic to Python, Cython packages that depend on numpy require numpy header files be installed before attempting to install the package you want. The numpy headers are not recognized unless numpy is installed in a seperate process that runs first. There are hacks for this issue, but I haven't gotten them to work. If you think binaries should be available for your platform, please let us know by opening an issue.\n\n#### Manual Installation\n\nThis can be desirable if you want to hack on CloudVolume itself.\n\n```bash\ngit clone git@github.com:seung-lab/cloud-volume.git\ncd cloud-volume\n\n# With virtualenvwrapper\nmkvirtualenv cv\nworkon cv\n# With only virtualenv\nvirtualenv venv\nsource venv/bin/activate\n\nsudo apt-get install g++ python3-dev # python-dev if you're on python2\npip install numpy # additional step needed for accelerated compressed_segmentation and fpzip\npip install -e . # without optional dependencies\npip install -e .[all_viewers] # with e.g. the all_viewers optional dependency\n```\n\n### Credentials\n\nBy default, CloudVolume's configuration and cache files are stored in `$HOME/.cloudvolume` (or in `$HOME/.cloudfiles` since we use CloudFiles for the backend). You can configure where CloudVolume looks for these files with the environment variable `$CLOUD_VOLUME_DIR`.  \n\nCredentials are stored in `$CLOUD_VOLUME_DIR/secrets`. You'll need credentials only for the services you'll use. If you plan to use the local filesystem, you won't need any. For Google Storage ([setup instructions here](https://github.com/seung-lab/cloud-volume/wiki/Setting-up-Google-Cloud-Storage)), default account credentials will be used if available and no service account is provided.\n\nIf neither of those two conditions apply, you need a service account credential. If you have your credentials handy, you can provide them like so as a dict, JSON string, or a bare token if the service will accept that.\n\n```python\ncv = CloudVolume(..., secrets=...)\n```\n\nHowever, it may be simpler to save your credential to disk so you don't have to always provide it. `google-secret.json` is a service account credential for Google Storage, `aws-secret.json` is a service account for S3, etc. You can support multiple projects at once by prefixing the bucket you are planning to access to the credential filename. `google-secret.json` will be your defaut service account, but if you also want to also access bucket ABC, you can provide `ABC-google-secret.json` and you'll have simultaneous access to your ordinary buckets and ABC. The secondary credentials are accessed on the basis of the bucket name, not the project name.\n\n```bash\nmkdir -p ~/.cloudvolume/secrets/\nmv aws-secret.json ~/.cloudvolume/secrets/ # needed for Amazon\nmv google-secret.json ~/.cloudvolume/secrets/ # needed for Google\nmv boss-secret.json ~/.cloudvolume/secrets/ # needed for the BOSS\nmv matrix-secret.json ~/.cloudvolume/secrets/ # needed for Matrix\nmv tigerdata-secret.json ~/.cloudvolume/secrets/ # needed for Tigerdata\n```\n\n#### `aws-secret.json` and `matrix-secret.json`\n\nCreate an [IAM user service account](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_users.html) that can read, write, and delete objects from at least one bucket.\n\n```json\n{\n\t\"AWS_ACCESS_KEY_ID\": \"$MY_AWS_ACCESS_KEY_ID\",\n\t\"AWS_SECRET_ACCESS_KEY\": \"$MY_SECRET_ACCESS_TOKEN\"\n}\n```\n\n#### `google-secret.json`\n\nYou can create the `google-secret.json` file [here](https://console.cloud.google.com/iam-admin/serviceaccounts). You don't need to manually fill in JSON by hand, the below example is provided to show you what the end result should look like. You should be able to read, write, and delete objects from at least one bucket.\n\n```json\n{\n  \"type\": \"service_account\",\n  \"project_id\": \"$YOUR_GOOGLE_PROJECT_ID\",\n  \"private_key_id\": \"...\",\n  \"private_key\": \"...\",\n  \"client_email\": \"...\",\n  \"client_id\": \"...\",\n  \"auth_uri\": \"https://accounts.google.com/o/oauth2/auth\",\n  \"token_uri\": \"https://accounts.google.com/o/oauth2/token\",\n  \"auth_provider_x509_cert_url\": \"https://www.googleapis.com/oauth2/v1/certs\",\n  \"client_x509_cert_url\": \"\"\n}\n```\n\n#### `cave-secret.json`\n\n*Note: used to be called chunkedgraph-secret.json. This is still supported but deprecated.*\n\nIf you have a token from Graphene/Chunkedgraph server, create the `cave-secret.json` file as shown in the example below. You may also pass the token to `CloudVolume(..., secrets=token)`.\n\n```json\n{\n  \"token\": \"<your_token>\"\n}\n```\n\nNote that to take advantage of multiple credential files, prepend the fully qualified domain name (FQDN) of the server instead of the bucket for GCS and S3. For example, `sudomain.domain.com-cave-secret.json`.\n\n## Usage\n\nCloudVolume supports reading and writing to Neuroglancer data layers on Amazon S3, Google Storage, The BOSS, and the local file system.\n\nSupported URLs are of the forms:\n\n`$FORMAT://$PROTOCOL://$BUCKET/$DATASET/$LAYER`\n\nThe format or protocol fields may be omitted where required. In the case of the precomputed format, the format specifier is optional.\n\n| Format      | Protocols                                    | Default | Example                                |\n|-------------|----------------------------------------------|---------|----------------------------------------|\n| precomputed | gs, s3, http, https, file, matrix, tigerdata | Yes     | gs://mybucket/dataset/layer            |\n| graphene    | gs, s3, http, https, file, matrix, tigerdata |         | graphene://gs://mybucket/dataset/layer |\n| boss        | N/A                                          |         | boss://collection/experiment/channel   |\n| n5          | gs, s3, http, https, file, matrix, tigerdata |         | n5://gs://mybucket/dataset/layer       |\n\n### Supported Formats\n\n* precomputed: Neuroglancer's native format. ([specification](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed))\n* graphene: Precomputed based format used by the PyChunkGraph server.\n* boss: The BOSS (https://docs.theboss.io/docs)\n* n5: Not HDF5 (https://github.com/saalfeldlab/n5) Read-only support. Supports raw, gzip, bz2, and xz but not lz4 compression. mode 0 datasets only.\n\n### Supported Protocols\n\n* gs:   Google Storage\n* s3:   Amazon S3\n* http(s): (read-only) Ordinary Web Servers\n* file: Local File System (absolute path)\n* matrix: Princeton Internal System (run in large part by Seung Lab)\n* tigerdata: Princeton Internal System (run by Princeton OIT)\n\nCloudVolume also supports [alternative s3 aliases](https://github.com/seung-lab/cloud-files#alias-for-alternative-s3-endpoints) via CloudFiles.\n\n\n### `info` Files - New Dataset\n\nNeuroglancer relies on an [`info`](https://github.com/google/neuroglancer/tree/master/src/neuroglancer/datasource/precomputed#info-json-file-specification) file located at the root of a dataset layer to tell it how to compute file locations and interpret the data in each file. CloudVolume piggy-backs on this functionality.\n\nIn the below example, assume you are creating a new segmentation volume from a 3d numpy array \"rawdata\". Note Precomputed stores data in Fortran (column major, aka CZYX) order. You should do a small test to see if the image is written transposed. You can fix this by uploading `rawdata.T`. A more detailed example for uploading a local volume [is located here](https://github.com/seung-lab/cloud-volume/wiki/Example-Single-Machine-Dataset-Upload).\n\n```python3\nfrom cloudvolume import CloudVolume\n\ninfo = CloudVolume.create_new_info(\n    num_channels    = 1,\n    layer_type      = 'segmentation',\n    data_type       = 'uint64', # Channel images might be 'uint8'\n    # raw, png, jpeg, compressed_segmentation, fpzip, kempressed, zfpc, compresso, crackle\n    encoding        = 'raw', \n    resolution      = [4, 4, 40], # Voxel scaling, units are in nanometers\n    voxel_offset    = [0, 0, 0], # x,y,z offset in voxels from the origin\n    mesh            = 'mesh',\n    # Pick a convenient size for your underlying chunk representation\n    # Powers of two are recommended, doesn't need to cover image exactly\n    chunk_size      = [ 512, 512, 16 ], # units are voxels\n    volume_size     = [ 250000, 250000, 25000 ], # e.g. a cubic millimeter dataset\n)\nvol = CloudVolume(cfg.path, info=info)\nvol.commit_info()\nvol[cfg.x: cfg.x + cfg.length, cfg.y:cfg.y + cfg.length, cfg.z: cfg.z + cfg.length] = rawdata[:,:,:]\n```\n| Encoding                | Image Type                 | Lossless | Neuroglancer Viewable | Description                                                                              |\n|-------------------------|----------------------------|----------|-------------|------------------------------------------------------------------------------------------|\n| raw                     | Any                        | Y        | Y           | Serialized numpy arrays.                                                                 |\n| png                     | Image                      | Y        | Y           | Multiple slices stiched into a single PNG.                                               |\n| jpeg                    | Image                      | N        | Y           | Multiple slices stiched into a single JPEG.                                              |\n| compressed_segmentation | Segmentation               | Y        | Y           | Renumbered numpy arrays to reduce data width. Also used by Neuroglancer internally.      |\n| compresso               | Segmentation               | Y        | Y           | Lossless high compression algorithm for connectomics segmentation.                       |\n| crackle                 | Segmentation               | Y        | Y*           | Lossless high compression algorithm for connectomics segmentation.                       |\n| fpzip                   | Floating Point             | Y        | Y*           | Takes advantage of IEEE 754 structure + L1 Lorenzo predictor to get higher compression.  |\n| kempressed              | Anisotropic Z Floating Point | N**      | Y*           | Adds manipulations on top of fpzip to achieve higher compression.                        |\n| zfpc                    | Alignment Vector Fields    | N***     | Y*          | zfp stream container.                        |\n\n\\* Not integrated into official Neuroglancer yet, but available on a [fork](https://allcodecs-dot-neuromancer-seung-import.appspot.com/) which can be seen on Github [here](https://github.com/william-silversmith/neuroglancer/tree/wms_combined_codecs).\n\\*\\* Lossless if your data can handle adding and then subtracting 2.\n\\*\\*\\* Lossless by default, but you probably want to use the lossy mode.\n\nNote on `compressed_segmentation`: To use, make sure `compressed_segmentation_block_size` is specified (usually `[8,8,8]`. This field will appear in the `info` file in the relevant scale.\n\nNote on `zfpc`: To configure, use the fields `zfpc_rate`, `zfpc_precision`, `zfpc_tolerance`, `zfpc_correlated_dims` in the relevant scale of the `info` file.\n\n\n### Examples\n\n```python\n# Basic Examples\nvol = CloudVolume('gs://mybucket/retina/image')\nvol = CloudVolume('gs://mybucket/retina/image', secrets=token, dict or json)\nvol = CloudVolume('gs://bucket/dataset/channel', mip=0, bounded=True, fill_missing=False)\nvol = CloudVolume('gs://bucket/dataset/channel', mip=[ 8, 8, 40 ], bounded=True, fill_missing=False) # set mip at this resolution\nvol = CloudVolume('gs://bucket/datasset/channel', info=info) # New info file from scratch\nimage = vol[:,:,:] # Download the entire image stack into a numpy array\nimage = vol.download(bbox, mip=2, renumber=True) # download w/ smaller dtype\nimage = vol.download(bbox, mip=2, label=777) # download binary image for label\nuniq = vol.unique(bbox, mip=0) # efficient extraction of unique labels\nlisting = vol.exists( np.s_[0:64, 0:128, 0:64] ) # get a report on which chunks actually exist\nexists = vol.image.has_data(mip=0) # boolean check to see if any data is there\nlisting = vol.delete( np.s_[0:64, 0:128, 0:64] ) # delete this region (bbox must be chunk aligned)\nvol[64:128, 64:128, 64:128] = image # Write a 64^3 image to the volume\nimg = vol.download_point( (x,y,z), size=256, mip=3 ) # download region around (mip 0) x,y,z at mip 3\npts = vol.scattered_points([ (x1,y1,z1), (x2,y2,z2) ]) # download voxel labels located at indicated points\n# download image files without decompressing or rendering them. Good for caching!\nfiles = vol.download_files(bbox, mip, decompress=False)\n\n# creates an anonymous in-memory CloudVolume that \n# will self-clean when the reference count drops to zero. \n# Store compressed images in memory for quick access!\nmem_vol = vol.image.memory_cutout(bbox, mip=1, encoding=\"compresso\")\n\n# Server\nvol.viewer() # launches neuroglancer compatible web server on http://localhost:1337\n\n# Microviewer (outdated, see https://github.com/seung-lab/microviewer/)\nimg = vol[64:1028, 64:1028, 64:128]\nimg.viewer() # launches web viewer on http://localhost:8080\n\n# Meshes\nvol.mesh.save(12345) # save 12345 as ./12345.ply on disk\nvol.mesh.save([12345, 12346, 12347]) # merge three segments into one file\nvol.mesh.save(12345, file_format='obj') # 'ply' and 'obj' are both supported\nvol.mesh.get(12345) # return the mesh as vertices and faces instead of writing to disk\nvol.mesh.get([ 12345, 12346 ]) # return these two segids fused into a single mesh\nvol.mesh.get([ 12345, 12346 ], fuse=False) # return { 12345: mesh, 12346: mesh }\nvol.mesh.put(meshes) # works for unsharded legacy only\nvol.mesh.delete(segids) # works for unsharded meshes only\n\nmesh.viewer() # Opens GUI. Requires vtk.\n\n# Skeletons\nskel = vol.skeleton.get(12345)\nvol.skeleton.upload_raw(segid, skel.vertices, skel.edges, skel.radii, skel.vertex_types)\nvol.skeleton.upload(skel)\n\n# specified in nm, only available for datasets with a generated index\nskels = vol.skeleton.get_by_bbox( Bbox( (0,0,0), (500, 500, 500) ) )\nvol.skeleton.spatial_index # None if not available\n\nskel.empty() # boolean\n\nbytes = skel.encode() # encode to Precomputed format (bytes)\nskel = Skeleton.decode(bytes) # decode from PrecomputedFormat\n\nskel = skel.crop(slices or bbox) # eliminate vertices and edges outside bbox\nskel = skel.consolidate() # eliminate duplicate vertices and edges\nskel3 = skel.merge(skel2) # merge two skeletons into one\nskel = skel.clone() # create copy\nskel = Skeleton.from_swc(swcstr) # decode an SWC file\nskel_str = skel.to_swc() # convert to SWC file in string representation\nskel.viewer() # Opens GUI. Requires matplotlib\n\nskel.cable_length() # sum of all edge lengths\nskel = skel.downsample(2) # reduce size of skeleton by factor of 2\n\nskel1 == skel2 # check if contents of internal arrays match\nSkeleton.equivalent(skel1, skel2) # ...even if there are differences like differently numbered edges\n\n# Parallel Operation\nvol = CloudVolume('gs://mybucket/retina/image', parallel=True) # Use all cores\nvol.parallel = 4 # e.g. any number > 1, use this many cores\ndata = vol[:] # uses shared memory to coordinate processes under the hood\n\n# Shared Memory Output (can be used by other processes)\nvol = CloudVolume(...)\n# data backed by a shared memory buffer\n# location is optional (defaults to vol.shared_memory_id)\ndata = vol.download_to_shared_memory(np.s_[:], location='some-example')\nvol.unlink_shared_memory() # delete the shared memory associated with this cloudvolume\nvol.shared_memory_id # get/set the default shared memory location for this instance\n\n# Shared Memory Upload\nvol = CloudVolume(...)\nvol.upload_from_shared_memory('my-shared-memory-id', # do not prefix with /dev/shm\n    bbox=Bbox( (0,0,0), (10000, 7500, 64) ))\n\n# Download or Upload directly with Files\n# The files must be in Precomputed raw format.\nvol.download_to_file('/path/to/file', bbox=Bbox(...), mip=0) # bbox is the download region\nvol.upload_from_file('/path/to/file', bbox=Bbox(...), mip=0) # bbox is the region it represents\n\n# Transfer w/o Excess Memory Allocation\nvol = CloudVolume(...)\n# single core, send all of vol to destination, no painting memory\n# you can also transcode the image encoding and compression type\nvol.transfer_to('gs://bucket/dataset/layer', vol.bounds)\n\n# Caching, default located at $HOME/.cloudvolume/cache/$PROTOCOL/$BUCKET/$DATASET/$LAYER/$RESOLUTION\n# You can also set the cache location using\n# cache=str or with environment variable CLOUD_VOLUME_CACHE_DIR\nvol = CloudVolume('gs://mybucket/retina/image', cache=True) # Basic Example\nimage = vol[0:10,0:10,0:10] # Download partial image and cache\nvol[0:10,0:10,0:10] = image # Upload partial image and cache\n\n# Resizing and clearing the LRU in-memory cache\nvol = CloudVolume(..., lru_bytes=num_bytes) # >= 0, 0 means disabled\nvol.image.lru.resize(num_bytes) # same\nvol.image.lru.clear()\nlen(vol.image.lru) # number of items in lru\nvol.image.lru.nbytes # size in bytes (not counting LRU structures, nor recursive)\nvol.image.lru.items() # etc, also functions as a dict\n\n# Evaluating the on-disk Cache\nvol.cache.list() # list files in cache at this mip level\nvol.cache.list(mip=1) # list files in cache at mip 1\nvol.cache.list_meshes()\nvol.cache.list_skeletons()\nvol.cache.num_files() # number of files at this mip level\nvol.cache.num_bytes(all_mips=True) # Return num files for each mip level in a list\nvol.cache.num_bytes() # number of bytes taken up by files, size on disk can be bigger\nvol.cache.num_bytes(all_mips=True) # Return num bytes for each mip level in a list\n\nvol.cache.enabled = True/False # Turn the cache on/off\nvol.cache.path = Str # set the cache location\nvol.cache.compress = None/True/False # None: Link to cloud setting, Boolean: Force cache to compressed (True) or uncompressed (False)\n\n# Deleting Cache\nvol.cache.flush() # Delete local cache for this layer at this mip level\nvol.cache.flush(preserve=Bbox(...)) # Same, but preserve cache in a region of space\nvol.cache.flush_region(region=Bbox(...), mips=[...]) # Delete the cached files in this region at these mip levels (default all mips)\nvol.cache.flush_info()\nvol.cache.flush_provenance()\n\n# Using Green Threads\nimport gevent.monkey\ngevent.monkey.patch_all(thread=False)\n\ncv = CloudVolume(..., green_threads=True)\nimg = cv[...] # now green threads will be used\n\n# Dask Interface (requires dask installation)\narr = cv.to_dask()\narr = cloudvolume.dask.from_cloudvolume(cloudpath) # same as to_dask\nres = cloudvolume.dask.to_cloudvolume(arr, cloudpath, compute=bool, return_store=bool)\n```\n\n### CloudVolume Constructor\n\n```python3\nCloudVolume(\n    cloudpath:str, mip:int=0, bounded:bool=True, \n    autocrop:bool=False, fill_missing:bool=False, cache:CacheType=False, \n    compress_cache:CompressType=None, cdn_cache:bool=True, \n    progress:bool=INTERACTIVE, info:dict=None, provenance:dict=None,\n    compress:CompressType=None, compress_level:Optional[int]=None, \n    non_aligned_writes:bool=False, parallel:ParallelType=1, delete_black_uploads:bool=False, \n    background_color:int=0, green_threads:bool=False, use_https:bool=False,\n    max_redirects:int=10, mesh_dir:Optional[str]=None, skel_dir:Optional[str]=None, \n    agglomerate:bool=False, secrets:SecretsType=None, \n    spatial_index_db:Optional[str]=None, lru_bytes:int = 0,\n    cache_locking:bool = True\n)\n```\n\n\n*      agglomerate: (bool, graphene only) sets the default mode for downloading\n        images to agglomerated (True) vs watershed (False).\n*      autocrop: (bool) If the specified retrieval bounding box exceeds the\n          volume bounds, process only the area contained inside the volume. \n          This can be useful way to ensure that you are staying inside the \n          bounds when `bounded=False`.\n*      background_color: (number) Specifies what the \"background value\" of the\n        volume is (traditionally 0). This is mainly for changing the behavior\n        of delete_black_uploads.\n*      bounded: (bool) If a region outside of volume bounds is accessed:\n          True: Throw an error\n          False: Allow accessing the region. If no files are present, an error \n              will still be thrown. Consider combining this option with \n              `fill_missing=True`. However, this can be dangrous as it allows\n              missing files and potentially network errors to be intepreted as \n              zeros.\n*      cache: (bool or str) Store downs and uploads in a cache on disk\n            and preferentially read from it before redownloading.\n          - falsey value: no caching will occur.\n          - True: cache will be located in a standard location.\n          - non-empty string: cache is located at this file path\n\n          After initialization, you can adjust this setting via:\n          `cv.cache.enabled = ...` which accepts the same values.\n\n          Note: This cache is totally separate from the LRU controlled by \n          lru_bytes.\n*      cache_locking: (bool) The local cache will use file locks via fasteners to prevent issues with multi-process cache access. If this is not a concern, performance can be slightly improved by setting this to False. This uses CloudFiles' locking mechanism.\n*      cdn_cache: (int, bool, or str) Sets Cache-Control HTTP header on uploaded \n        image files. Most cloud providers perform some kind of caching. As of \n        this writing, Google defaults to 3600 seconds. Most of the time you'll \n        want to go with the default. \n        - int: number of seconds for cache to be considered fresh (max-age)\n        - bool: True: max-age=3600, False: no-cache\n        - str: set the header manually\n*      compress: (bool, str, None) pick which compression method to use.\n*          None: (default) gzip for raw arrays and no additional compression\n            for compressed_segmentation and fpzip.\n          bool: \n            True=gzip, \n            False=no compression, Overrides defaults\n          str: \n            'gzip': Extension so that we can add additional methods in the future \n                    like lz4 or zstd. \n            'br': Brotli compression, better compression rate than gzip\n            '': no compression (same as False).\n*      compress_level: (int, None) level for compression. Higher number results\n          in better compression but takes longer.\n        Defaults to 9 for gzip (ranges from 0 to 9).\n        Defaults to 5 for brotli (ranges from 0 to 11).\n*      compress_cache: (None or bool) If not None, override default compression \n          behavior for the cache.\n*      delete_black_uploads: (bool) If True, on uploading an entirely black chunk,\n          issue a DELETE request instead of a PUT. This can be useful for avoiding storing\n          tiny files in the region around an ROI. Some storage systems using erasure coding \n          don't do well with tiny file sizes.\n*      fill_missing: (bool) If a chunk file is unable to be fetched:\n          True: Use a block of zeros\n          False: Throw an error\n*      green_threads: (bool) Use green threads instead of preemptive threads. This\n        can result in higher download performance for some compression types. Preemptive\n        threads seem to reduce performance on multi-core machines that aren't densely\n        loaded as the CPython threads are assigned to multiple cores and the thrashing\n        + GIL reduces performance. You'll need to add the following code to the top\n        of your program to use green threads:\n\n            import gevent.monkey\n            gevent.monkey.patch_all(threads=False)\n*      lru_bytes: (int) number of bytes used to cache recently used image \n        tiles in memory. This is an in-memory cache and is completely separate from\n        the `cache` parameter that handles disk IO. Tiles are stripped over only their\n        second stage compression.\n*      info: (dict) In lieu of fetching a neuroglancer info file, use this one.\n          This is useful when creating new datasets and for repeatedly initializing\n          a new cloudvolume instance.\n*      max_redirects: (int) if > 0, allow up to this many redirects via info file 'redirect'\n          data fields. If <= 0, allow no redirections and access the current info file directly\n          without raising an error.\n*      mesh_dir: (str) if not None, override the info['mesh'] key before pulling the\n        mesh info file.\n*      mip: (int or iterable) Which level of downsampling to read and write from.\n          0 is the highest resolution. You can also specify the voxel resolution\n          like mip=[6,6,30] which will search for the appropriate mip level.\n*      non_aligned_writes: (bool) Enable non-aligned writes. Not multiprocessing \n          safe without careful design. When not enabled, a \n          cloudvolume.exceptions.AlignmentError is thrown for non-aligned writes. \n          \n          https://github.com/seung-lab/cloud-volume/wiki/Advanced-Topic:-Non-Aligned-Writes\n\n      parallel (int: 1, bool): Number of extra processes to launch, 1 means only \n          use the main process. If parallel is True use the number of CPUs \n          returned by multiprocessing.cpu_count(). When parallel > 1, shared\n          memory (Linux) or emulated shared memory via files (other platforms) \n          is used by the underlying download.\n*      progress: (bool) Show progress bars. \n          Defaults to True in interactive python, False in script execution mode.\n*      provenance: (string, dict) In lieu of fetching a provenance \n          file, use this one. \n*      secrets: (dict) provide per-instance authorization tokens. If not provided,\n        defaults to looking in .cloudvolume/secrets for necessary tokens.\n*      skel_dir: (str) if not None, override the info['skeletons'] key before \n        pulling the skeleton info file.\n*      spatial_index_db: (str) A path to an sqlite3 or mysql database that follows \n        the following uri schema. sqlite is assumed if no scheme is present in \n        the uri.\n          [sqlite://]filename.db\n          mysql://<username>:<password>@<host>:<port>/<db_name>\n\n        Igneous generated datasets include a JSON based spatial\n        database that tiles the dataset. This can be fast enough up to about 100 TVx\n        datasets. Above that, a proper database is required for efficient queries.\n        We provide multiple SQL database types that the index can be hosted on.\n*      use_https: (bool) maps gs:// and s3:// to their respective https paths. The \n        https paths hit a cached, read-only version of the data and may be faster.\n\n### CloudVolume Methods\n\nBetter documentation coming later, but for now, here's a summary of the most useful method calls. Use help(cloudvolume.CloudVolume.$method) for more info.\n\n* create_new_info (class method) - Helper function for creating info files for creating new data layers.\n* refresh_info - Repull the info file.\n* refresh_provenance - Repull the provenance file.\n* bbox_to_mip - Covert a bounding box or slice from one mip level to another.\n* slices_from_global_coords - *deprecated, why not use bbox_to_mip?* Find the CloudVolume slice from MIP 0 coordinates if you're on a different MIP. Often used in combination with neuroglancer.\n* reset_scales - Delete mips other than 0 in the info file. Does not autocommit.\n* add_scale - Generate a new mip level in the info property. Does not autocommit.\n* commit_info - Push the current info property into the cloud as a JSON file.\n* commit_provenance - Push the current provenance property into the cloud as a JSON file.\n* image - Access image operations directly.\n  * download - Download bounding boxes from a given mip level.\n  * upload - Upload images to bounding boxes at a given mip level.\n  * transfer_to - Transfer data without painting a container array to avoid out of memory errors.\n  * exists - Check which chunk files exist in a given bounding box.\n  * delete - Delete chunks in a given bounding box at a given mip level.\n* mesh - Access mesh operations\n\t* get - Download an object. Can merge multiple segmentids\n\t* save - Download an object and save it in `.obj` format. You can combine equivialences into a single object too.\n* skeleton - Access Skeletons\n  * get - Download an object.\n  * upload - Save a skeleton object to the cloud.\n* cache - Access cache operations\n\t* enabled - Boolean switch to enable/disable cache. If true, on reading, check local disk cache before downloading, and save downloaded chunks to cache. When writing, write to the cloud then save the chunks you wrote to cache. If false, bypass cache completely. The cache is located at `$HOME/.cloudvolume/cache`.\n\t* path - Property that shows the current filesystem path to the cache\n\t* list - List files in cache\n\t* num_files - Number of files in cache at this mip level , use all_mips=True to get them all\n\t* num_bytes - Return the number of bytes in cache at this mip level, all_mips=True to get them all\n\t* flush - Delete the cache at this mip level, preserve=Bbox/slice to save a spatial region\n\t* flush_region - Delete a spatial region at this mip level\n* exists - Generate a report on which chunks within a bounding box exist.\n* delete - Delete the chunks within this bounding box.\n* transfer_to - Transfer data from a bounding box to another data storage location. Does not allocate memory and transfers in blocks, so can transfer large volumes of data. May be less efficient than a dedicated tool like `gsutil` or `aws s3`.\n* unlink_shared_memory - Delete shared memory associated with this instance (`vol.shared_memory_id`)\n* generate_shared_memory_location - Create a new unique shared memory identifier string. No side effects.\n* download_to_shared_memory - Instead of using ordinary numpy memory allocations, download to shared memory.\n    Be careful, shared memory is like a file and doesn't disappear unless explicitly unlinked. (`vol.unlink_shared_memory()`)\n* upload_from_shared_memory - Upload from a given shared memory block without making a copy.\n* download_point - Download the region around this mip 0 coordinate at a given mip level.\n\n### CloudVolume Properties\n\nAccessed as `vol.$PROPERTY` like `vol.mip`. Parens next to each property mean (data type:default, writability). (r) means read only, (w) means write only, (rw) means read/write.\n\n* mip (uint:0, rw) - Read from and write to this mip level (0 is highest res). Each additional increment in the number is typically a 2x reduction in resolution.\n* bounded (bool:True, rw) - If a region outside of volume bounds is accessed throw an error if True or Fill the region with black (useful for e.g. marching cubes's 1px boundary) if False.\n* autocrop (bool:False, rw) - If bounded is False and this option is True, automatically crop requested uploads and downloads to the volume boundary.\n* fill_missing (bool:False, rw) - If a file inside volume bounds is unable to be fetched use a block of zeros if True, else throw an error.\n* delete_black_uploads (bool:False, rw) - If True, issue a DELETE http request instead of a PUT when an individual uploaded chunk is all zeros.\n* info (dict, rw) - Python dict representation of Neuroglancer info JSON file. You must call `vol.commit_info()` to save your changes to storage.\n* provenance (dict-like, rw) - Data layer provenance file representation. You must call `vol.commit_provenance()` to save your changes to storage.\n* available_mips (list of ints, r) - Query which mip levels are defined for reading and writing.\n* dataset_name (str, rw) - Which dataset (e.g. test_v0, snemi3d_v0) on S3, GS, or FS you're reading and writing to. Known as an \"experiment\" in BOSS terminology. Writing to this property triggers an info refresh.\n* layer (str, rw) - Which data layer (e.g. image, segmentation) on S3, GS, or FS you're reading and writing to. Known as a \"channel\" in BOSS terminology. Writing to this property triggers an info refresh.\n* base_cloudpath (str, r) - The cloud path to the dataset e.g. s3://bucket/dataset/\n* layer_cloudpath (str, r) - The cloud path to the data layer e.g. gs://bucket/dataset/image\n* info_cloudpath (str, r) - Generate the cloud path to this data layer's info file.\n* scales (dict, r) - Shortcut to the 'scales' property of the info object\n* scale (dict, rw)* - Shortcut to the working scale of the current mip level\n* shape (Vec4, r)* - Like numpy.ndarray.shape for the entire data layer.\n* volume_size (Vec3, r)* - Like shape, but omits channel (x,y,z only).\n* num_channels (int, r) - The number of channels, the last element of shape.\n* layer_type (str, r) - The neuroglancer info type, 'image' or 'segmentation'.\n* dtype (str, r) - The info data_type of the volume, e.g. uint8, uint32, etc. Similar to numpy.ndarray.dtype.\n* encoding (str, r) - The neuroglancer info encoding. e.g. 'raw', 'jpeg', 'npz'\n* resolution (Vec3, r)* - The 3D physical resolution of a voxel in nanometers at the working mip level.\n* downsample_ratio (Vec3, r) - Ratio of the current resolution to the highest resolution mip available.\n* chunk_size (Vec3, r)* - Size of the underlying chunks that constitute the volume in storage. e.g. Vec(64, 64, 64)\n* key (str, r)* - The 'directory' we're accessing the current working mip level from within the data layer. e.g. '6_6_30'\n* bounds (Bbox, r)* - A Bbox object that represents the bounds of the entire volume.\n* shared_memory_id (str, rw) - Shared memory location used for parallel operation or for output.\n\n\\* These properties can also be accessed with a function named like `vol.mip_$PROPERTY($MIP)`. By default they return the current mip level assigned to the CloudVolume, but any mip level can be accessed via the corresponding `mip_` function. Example: `vol.mip_resolution(2)` would return the resolution of mip 2.\n\n### VolumeCutout Functions\n\nWhen you download an image using CloudVolume it gives you a `VolumeCutout`. These are `numpy.ndarray` subclasses that support a few extra properties to help make book keeping easier. The major advantage is `save_images()` which can help you view your dataset as PNG slices.\n\n* `dataset_name` - The dataset this image came from.\n* `layer` - Which layer it came from.\n* `mip` - Which mip it came from\n* `layer_type` - \"image\" or \"segmentation\"\n* `bounds` - The bounding box of the cutout\n* `num_channels` - Alias for `vol.shape[3]`\n* `save_images()` - Save Z slice PNGs of the current image to `./saved_images` for manual inspection\n* `viewer()` - Start a local web server (http://localhost:8080) that can view small volumes interactively. This was recently changed from `view` as `view` is a useful numpy method.\n\n### Viewing a Precomputed Volume on Disk\n\nIf you have Precomputed volume onto local disk and would like to point neuroglancer to it:\n\n```python\nvol = CloudVolume(...)\nvol.viewer()\n```\n\nYou can then point any version of neuroglancer at it using `precomputed://http://localhost:1337/NAME_OF_LAYER`.\n\n### Microviewer\n\nCloudVolume includes a built-in dependency free viewer for 3D volumetric datasets smaller than about 2GB uncompressed. It supports bool, uint8, uint16, uint32, float32, and float64 numpy data types for both images and segmentation and can render a composite overlay of image and segmentation.\n\nYou can launch a viewer using the `.viewer()` method of a VolumeCutout object or by using the `view(...)` or `hyperview(...)` functions that come with the cloudvolume module. This launches a web server on `http://localhost:8080`. You can read more [on the wiki](https://github.com/seung-lab/cloud-volume/wiki/%CE%BCViewer).\n\n```python3\nfrom cloudvolume import CloudVolume, view, hyperview\n\nchannel_vol = CloudVolume(...)\nseg_vol = CloudVolume(...)\nimg = vol[...]\nseg = vol[...]\n\nimg.viewer() # works on VolumeCutouts\nseg.viewer() # segmentation type derived from info\nview(img) # alternative for arbitrary numpy arrays\nview(seg, segmentation=True)\nhyperview(img, seg) # img and seg shape must match\n\n>>> Viewer server listening to http://localhost:8080\n```\n\nThere are also seperate viewers for skeleton and mesh objects that can be invoked by calling `.viewer()` on either object. However, skeletons depend on `matplotlib` and meshes depend on `vtk` and OpenGL to function.\n\n```bash\npip install vtk matplotlib\n```\n\n## Python 2.7 End of Life\n\nPython 2.7 is no longer supported by CloudVolume. Updated versions of `pip` will download the last supported release 1.21.1. You can read more on the policy page: https://github.com/seung-lab/cloud-volume/wiki/Policy#python-27-end-of-life\n\n## Related Projects\n\n1. [Igneous](https://github.com/seung-lab/igneous): Computational pipeline for visualizing neuroglancer volumes.\n2. [CloudVolume.jl](https://github.com/seung-lab/CloudVolume.jl): CloudVolume in Julia\n3. [fpzip](https://github.com/seung-lab/fpzip): A Python Package for the C++ code by Lindstrom et al.\n4. [compressed_segmentation](https://github.com/seung-lab/compressedseg): A Python Package wrapping the code for the compressed_segmentation format developed by Jeremy Maitin-Shepard and Stephen Plaza.\n5. [Kimimaro](https://github.com/seung-lab/kimimaro): High performance skeletonization of densely labeled 3D volumes.\n6. [compresso](https://github.com/seung-lab/compresso): High lossless compression of connectomics segmentation. Algorithm by and code derived from Matejek et al.\n7. [zfpc](https://github.com/seung-lab/zfpc): Optimized zfp multi-stream container for alignment vector fields (and similar floating point data).\n8. [crackle](https://github.com/seung-lab/crackle): Lossless high compression of connectomics segmentation. (BETA)\n\n## Acknowledgments\n\nThank you to everyone that has contributed past or current to CloudVolume or the ecosystem it serves. We love you!  \n\nJeremy Maitin-Shepard created [Neuroglancer](https://github.com/google/neuroglancer) and defined the Precomputed format. Yann Leprince provided a [pure Python codec](https://github.com/HumanBrainProject/neuroglancer-scripts) for the compressed_segmentation format. Jeremy Maitin-Shepard and Stephen Plaza created C++ code defining the compression and decompression (respectively) protocol for [compressed_segmentation](https://github.com/janelia-flyem/compressedseg). Peter Lindstrom et al. created [the fpzip algorithm](https://computation.llnl.gov/projects/floating-point-compression), and contributed a C++ implementation and advice. Nico Kemnitz adapted our data to fpzip using the \"Kempression\" protocol (we named it, not him). Dan Bumbarger contributed code and information helpful for getting CloudVolume working on Windows. Fredrik Kihlander's [pure python implementation](https://github.com/wc-duck/pymmh3) of murmurhash3 and [Austin Appleby](https://github.com/aappleby/smhasher) developed murmurhash3 which is necessary for the sharded format. Ben Falk advocated for and did the bulk of the work on brotli compression. Some of the ideas in CloudVolume are based on work by Jingpeng Wu in [BigArrays.jl](https://github.com/seung-lab/BigArrays.jl).  Sven Dorkenwald, Manuel Castro, and Akhilesh Halageri contributed advice and code towards implementing the graphene interface. Oluwaseun Ogedengbe contributed documentation for the sharded format. Eric Perlman wrote the reader for Neuroglancer Multi-LOD meshes. Ignacio Tartavull and William Silversmith wrote the initial version of CloudVolume.\n",
    "bugtrack_url": null,
    "license": "License :: OSI Approved :: BSD License",
    "summary": "A serverless client for reading and writing Neuroglancer Precomputed volumes both locally and on cloud services.",
    "version": "8.32.1",
    "project_urls": {
        "Homepage": "https://github.com/seung-lab/cloud-volume/"
    },
    "split_keywords": [
        "neuroglancer",
        "volumetric-data",
        "numpy",
        "connectomics",
        "microscopy",
        "image-processing",
        "biomedical-image-processing",
        "s3",
        "gcs",
        "mesh",
        "meshes",
        "skeleton",
        "skeletons"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "85b40faa5536d75d775a1522141ea8e1864bfcaddd8738be451272558d5fca9d",
                "md5": "30593d305ff581b2a512b3ba8fecee5e",
                "sha256": "d4a25154a2da75e4d58c2c597eb788e0b8f6674646da0fc68d8f1ed1d4e13abe"
            },
            "downloads": -1,
            "filename": "cloud_volume-8.32.1-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "30593d305ff581b2a512b3ba8fecee5e",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": "~=3.6",
            "size": 242420,
            "upload_time": "2024-04-10T06:08:41",
            "upload_time_iso_8601": "2024-04-10T06:08:41.634543Z",
            "url": "https://files.pythonhosted.org/packages/85/b4/0faa5536d75d775a1522141ea8e1864bfcaddd8738be451272558d5fca9d/cloud_volume-8.32.1-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "8ea7b59826cf80ff1f4507ccee39571aec004ae78a07d62680efdfe288a3987d",
                "md5": "63555bb6b4689ff50bee861e7fa23ac1",
                "sha256": "c9f1041b2f8d9c423d728bffef431d7bf265494d2e6bf740e5f39caae6f2c7e5"
            },
            "downloads": -1,
            "filename": "cloud-volume-8.32.1.tar.gz",
            "has_sig": false,
            "md5_digest": "63555bb6b4689ff50bee861e7fa23ac1",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": "~=3.6",
            "size": 280333,
            "upload_time": "2024-04-10T06:08:46",
            "upload_time_iso_8601": "2024-04-10T06:08:46.896255Z",
            "url": "https://files.pythonhosted.org/packages/8e/a7/b59826cf80ff1f4507ccee39571aec004ae78a07d62680efdfe288a3987d/cloud-volume-8.32.1.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-04-10 06:08:46",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "seung-lab",
    "github_project": "cloud-volume",
    "travis_ci": true,
    "coveralls": false,
    "github_actions": false,
    "requirements": [
        {
            "name": "boto3",
            "specs": [
                [
                    ">=",
                    "1.4.7"
                ]
            ]
        },
        {
            "name": "chardet",
            "specs": [
                [
                    ">=",
                    "3.0.4"
                ]
            ]
        },
        {
            "name": "cloud-files",
            "specs": [
                [
                    "<",
                    "5.0.0"
                ],
                [
                    ">=",
                    "4.24.0"
                ]
            ]
        },
        {
            "name": "compressed-segmentation",
            "specs": [
                [
                    ">=",
                    "2.1.1"
                ]
            ]
        },
        {
            "name": "compresso",
            "specs": [
                [
                    ">=",
                    "3.0.0"
                ]
            ]
        },
        {
            "name": "crackle-codec",
            "specs": [
                [
                    "<",
                    "2.0.0"
                ],
                [
                    ">=",
                    "0.8.0"
                ]
            ]
        },
        {
            "name": "DracoPy",
            "specs": [
                [
                    ">=",
                    "1.0.0"
                ],
                [
                    "<",
                    "2.0.0"
                ]
            ]
        },
        {
            "name": "fastremap",
            "specs": [
                [
                    ">=",
                    "1.14.0"
                ]
            ]
        },
        {
            "name": "fpzip",
            "specs": [
                [
                    "<",
                    "2.0.0"
                ],
                [
                    ">=",
                    "1.2.0"
                ]
            ]
        },
        {
            "name": "gevent",
            "specs": []
        },
        {
            "name": "google-auth",
            "specs": [
                [
                    ">=",
                    "1.10.0"
                ]
            ]
        },
        {
            "name": "google-cloud-core",
            "specs": [
                [
                    ">=",
                    "1.1.0"
                ]
            ]
        },
        {
            "name": "google-cloud-storage",
            "specs": [
                [
                    ">=",
                    "1.24.1"
                ]
            ]
        },
        {
            "name": "json5",
            "specs": []
        },
        {
            "name": "jsonschema",
            "specs": [
                [
                    "!=",
                    "4.0"
                ]
            ]
        },
        {
            "name": "numpy",
            "specs": [
                [
                    ">=",
                    "1.13.3"
                ]
            ]
        },
        {
            "name": "networkx",
            "specs": []
        },
        {
            "name": "python-jsonschema-objects",
            "specs": [
                [
                    ">=",
                    "0.3.3"
                ]
            ]
        },
        {
            "name": "pathos",
            "specs": []
        },
        {
            "name": "protobuf",
            "specs": [
                [
                    ">=",
                    "3.3.0"
                ]
            ]
        },
        {
            "name": "pyspng-seunglab",
            "specs": [
                [
                    ">=",
                    "1.0.0"
                ]
            ]
        },
        {
            "name": "python-dateutil",
            "specs": []
        },
        {
            "name": "requests",
            "specs": [
                [
                    ">=",
                    "2.22.0"
                ]
            ]
        },
        {
            "name": "psutil",
            "specs": []
        },
        {
            "name": "pysimdjson",
            "specs": [
                [
                    ">=",
                    "3.1.1"
                ]
            ]
        },
        {
            "name": "simplejpeg",
            "specs": []
        },
        {
            "name": "six",
            "specs": [
                [
                    ">=",
                    "1.10.0"
                ]
            ]
        },
        {
            "name": "tenacity",
            "specs": [
                [
                    ">=",
                    "4.10.0"
                ]
            ]
        },
        {
            "name": "tqdm",
            "specs": []
        },
        {
            "name": "urllib3",
            "specs": [
                [
                    ">=",
                    "1.25.7"
                ]
            ]
        },
        {
            "name": "zfpc",
            "specs": []
        }
    ],
    "tox": true,
    "lcname": "cloud-volume"
}
        
Elapsed time: 0.22662s