multi-model-server-gpu


Namemulti-model-server-gpu JSON
Version 0.0.2 PyPI version JSON
download
home_pagehttps://github.com/deathstarenterprise/multi-model-server-gpu
Summary(Altered for multiprocessing GPU Inference) Multi Model Server is a tool for serving neural net models for inference
upload_time2024-09-09 19:18:04
maintainerNone
docs_urlNone
authorTrinity team + Blake Donahoo
requires_pythonNone
licenseApache License Version 2.0
keywords multi model server serving deep learning inference ai (altered for multiprocessing gpu inference)
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage
            Project Description
===================

*This version of multi-model-server has been forked from version 1.1.11 and altered for multithreading GPU inference in scenarios where preload_model=True. Using the traditional multiprocessing library resulted in shortcomings when attempting to share models where GPU memory was allocated. This fork is an attempt to solve that. All other work within this library is to the credit of the original authors. ~Blake@duke.ai*

===================

Multi Model Server (MMS) is a flexible and easy to use tool for
serving deep learning models exported from `MXNet <http://mxnet.io/>`__
or the Open Neural Network Exchange (`ONNX <http://onnx.ai/>`__).

Use the MMS Server CLI, or the pre-configured Docker images, to start a
service that sets up HTTP endpoints to handle model inference requests.

Detailed documentation and examples are provided in the `docs
folder <https://github.com/awslabs/multi-model-server/blob/master/docs/README.md>`__.

Prerequisites
-------------

* **java 8**: Required. MMS use java to serve HTTP requests. You must install java 8 (or later) and make sure java is on available in $PATH environment variable *before* installing MMS. If you have multiple java installed, you can use $JAVA_HOME environment vairable to control which java to use.
* **mxnet**: `mxnet` will not be installed by default with MMS 1.0 any more. You have to install it manually if you use MxNet.

For ubuntu:
::

    sudo apt-get install openjdk-8-jre-headless


For centos
::

    sudo yum install java-1.8.0-openjdk


For Mac:
::

    brew tap caskroom/versions
    brew update
    brew cask install java8


Install MxNet:
::

    pip install mxnet

MXNet offers MKL pip packages that will be much faster when running on Intel hardware.
To install mkl package for CPU:
::

    pip install mxnet-mkl

or for GPU instance:

::

    pip install mxnet-cu92mkl


Installation
------------

::

    pip install multi-model-server

Development
-----------

We welcome new contributors of all experience levels. For information on
how to install MMS for development, refer to the `MMS
docs <https://github.com/awslabs/multi-model-server/blob/master/docs/install.md>`__.

Important links
---------------

-  `Official source code
   repo <https://github.com/awslabs/multi-model-server>`__
-  `Download
   releases <https://pypi.org/project/multi-model-server/#files>`__
-  `Issue
   tracker <https://github.com/awslabs/multi-model-server/issues>`__

Source code
-----------

You can check the latest source code as follows:

::

    git clone https://github.com/awslabs/multi-model-server.git

Testing
-------

After installation, try out the MMS Quickstart for

- `Serving a Model <https://github.com/awslabs/multi-model-server/blob/master/README.md#serve-a-model>`__
- `Create a Model Archive <https://github.com/awslabs/multi-model-server/blob/master/README.md#model-archive>`__.

Help and Support
----------------

-  `Documentation <https://github.com/awslabs/multi-model-server/blob/master/docs/README.md>`__
-  `Forum <https://discuss.mxnet.io/latest>`__

Citation
--------

If you use MMS in a publication or project, please cite MMS:
https://github.com/awslabs/multi-model-server

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/deathstarenterprise/multi-model-server-gpu",
    "name": "multi-model-server-gpu",
    "maintainer": null,
    "docs_url": null,
    "requires_python": null,
    "maintainer_email": null,
    "keywords": "Multi Model Server Serving Deep Learning Inference AI (Altered for multiprocessing GPU inference)",
    "author": "Trinity team + Blake Donahoo",
    "author_email": "noreply@amazon.com, blake@duke.ai",
    "download_url": "https://files.pythonhosted.org/packages/aa/b2/9f689a4e39adf85ea157b2bb1bd78a0202bb9c12b28945feb4c690386f8f/multi_model_server_gpu-0.0.2.tar.gz",
    "platform": null,
    "description": "Project Description\n===================\n\n*This version of multi-model-server has been forked from version 1.1.11 and altered for multithreading GPU inference in scenarios where preload_model=True. Using the traditional multiprocessing library resulted in shortcomings when attempting to share models where GPU memory was allocated. This fork is an attempt to solve that. All other work within this library is to the credit of the original authors. ~Blake@duke.ai*\n\n===================\n\nMulti Model Server (MMS) is a flexible and easy to use tool for\nserving deep learning models exported from `MXNet <http://mxnet.io/>`__\nor the Open Neural Network Exchange (`ONNX <http://onnx.ai/>`__).\n\nUse the MMS Server CLI, or the pre-configured Docker images, to start a\nservice that sets up HTTP endpoints to handle model inference requests.\n\nDetailed documentation and examples are provided in the `docs\nfolder <https://github.com/awslabs/multi-model-server/blob/master/docs/README.md>`__.\n\nPrerequisites\n-------------\n\n* **java 8**: Required. MMS use java to serve HTTP requests. You must install java 8 (or later) and make sure java is on available in $PATH environment variable *before* installing MMS. If you have multiple java installed, you can use $JAVA_HOME environment vairable to control which java to use.\n* **mxnet**: `mxnet` will not be installed by default with MMS 1.0 any more. You have to install it manually if you use MxNet.\n\nFor ubuntu:\n::\n\n    sudo apt-get install openjdk-8-jre-headless\n\n\nFor centos\n::\n\n    sudo yum install java-1.8.0-openjdk\n\n\nFor Mac:\n::\n\n    brew tap caskroom/versions\n    brew update\n    brew cask install java8\n\n\nInstall MxNet:\n::\n\n    pip install mxnet\n\nMXNet offers MKL pip packages that will be much faster when running on Intel hardware.\nTo install mkl package for CPU:\n::\n\n    pip install mxnet-mkl\n\nor for GPU instance:\n\n::\n\n    pip install mxnet-cu92mkl\n\n\nInstallation\n------------\n\n::\n\n    pip install multi-model-server\n\nDevelopment\n-----------\n\nWe welcome new contributors of all experience levels. For information on\nhow to install MMS for development, refer to the `MMS\ndocs <https://github.com/awslabs/multi-model-server/blob/master/docs/install.md>`__.\n\nImportant links\n---------------\n\n-  `Official source code\n   repo <https://github.com/awslabs/multi-model-server>`__\n-  `Download\n   releases <https://pypi.org/project/multi-model-server/#files>`__\n-  `Issue\n   tracker <https://github.com/awslabs/multi-model-server/issues>`__\n\nSource code\n-----------\n\nYou can check the latest source code as follows:\n\n::\n\n    git clone https://github.com/awslabs/multi-model-server.git\n\nTesting\n-------\n\nAfter installation, try out the MMS Quickstart for\n\n- `Serving a Model <https://github.com/awslabs/multi-model-server/blob/master/README.md#serve-a-model>`__\n- `Create a Model Archive <https://github.com/awslabs/multi-model-server/blob/master/README.md#model-archive>`__.\n\nHelp and Support\n----------------\n\n-  `Documentation <https://github.com/awslabs/multi-model-server/blob/master/docs/README.md>`__\n-  `Forum <https://discuss.mxnet.io/latest>`__\n\nCitation\n--------\n\nIf you use MMS in a publication or project, please cite MMS:\nhttps://github.com/awslabs/multi-model-server\n",
    "bugtrack_url": null,
    "license": "Apache License Version 2.0",
    "summary": "(Altered for multiprocessing GPU Inference) Multi Model Server is a tool for serving neural net models for inference",
    "version": "0.0.2",
    "project_urls": {
        "Homepage": "https://github.com/deathstarenterprise/multi-model-server-gpu"
    },
    "split_keywords": [
        "multi",
        "model",
        "server",
        "serving",
        "deep",
        "learning",
        "inference",
        "ai",
        "(altered",
        "for",
        "multiprocessing",
        "gpu",
        "inference)"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "c152dd74513dc5758b93b9d59539336948b82ddba243d9272a352aee13c6928f",
                "md5": "43461c59eb74192223240be8e687045c",
                "sha256": "d167071905fe742ceb2c068d479672c15241a7eca55cca5e5beb8307e1a38922"
            },
            "downloads": -1,
            "filename": "multi_model_server_gpu-0.0.2-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "43461c59eb74192223240be8e687045c",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": null,
            "size": 53547,
            "upload_time": "2024-09-09T19:18:01",
            "upload_time_iso_8601": "2024-09-09T19:18:01.550592Z",
            "url": "https://files.pythonhosted.org/packages/c1/52/dd74513dc5758b93b9d59539336948b82ddba243d9272a352aee13c6928f/multi_model_server_gpu-0.0.2-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "aab29f689a4e39adf85ea157b2bb1bd78a0202bb9c12b28945feb4c690386f8f",
                "md5": "c6bcae09ea743a657d768873c22732e9",
                "sha256": "5ec6a0d168b32389c24159c3e01094010dac23029fa77b20a01ad703057d2368"
            },
            "downloads": -1,
            "filename": "multi_model_server_gpu-0.0.2.tar.gz",
            "has_sig": false,
            "md5_digest": "c6bcae09ea743a657d768873c22732e9",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": null,
            "size": 38723,
            "upload_time": "2024-09-09T19:18:04",
            "upload_time_iso_8601": "2024-09-09T19:18:04.566666Z",
            "url": "https://files.pythonhosted.org/packages/aa/b2/9f689a4e39adf85ea157b2bb1bd78a0202bb9c12b28945feb4c690386f8f/multi_model_server_gpu-0.0.2.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-09-09 19:18:04",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "deathstarenterprise",
    "github_project": "multi-model-server-gpu",
    "travis_ci": false,
    "coveralls": true,
    "github_actions": false,
    "circle": true,
    "lcname": "multi-model-server-gpu"
}
        
Elapsed time: 0.32135s