# FBGEMM_GPU
[![FBGEMM_GPU-CPU CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cpu.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cpu.yml)
[![FBGEMM_GPU-CUDA CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cuda.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cuda.yml)
[![FBGEMM_GPU-ROCm CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_rocm.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_rocm.yml)
FBGEMM_GPU (FBGEMM GPU Kernels Library) is a collection of high-performance
PyTorch GPU operator libraries for training and inference. The library provides
efficient table batched embedding bag, data layout transformation, and
quantization supports.
FBGEMM_GPU is currently tested with CUDA 12.1 and 11.8 in CI, and with PyTorch
packages (2.1+) that are built against those CUDA versions.
See the full [Documentation](https://pytorch.org/FBGEMM) for more information
on building, installing, and developing with FBGEMM_GPU, as well as the most
up-to-date support matrix for this library.
## Join the FBGEMM_GPU Community
For questions, support, news updates, or feature requests, please feel free to:
* File a ticket in [GitHub Issues](https://github.com/pytorch/FBGEMM/issues)
* Post a discussion in [GitHub Discussions](https://github.com/pytorch/FBGEMM/discussions)
* Reach out to us on the `#fbgemm` channel in [PyTorch Slack](https://bit.ly/ptslack)
For contributions, please see the [`CONTRIBUTING`](../CONTRIBUTING.md) file for
ways to help out.
## License
FBGEMM_GPU is BSD licensed, as found in the [`LICENSE`](../LICENSE) file.
Raw data
{
"_id": null,
"home_page": "https://github.com/pytorch/fbgemm",
"name": "fbgemm-gpu-nightly-genai",
"maintainer": null,
"docs_url": null,
"requires_python": null,
"maintainer_email": null,
"keywords": "PyTorch, Recommendation Models, High Performance Computing, GPU, CUDA",
"author": "FBGEMM Team",
"author_email": "packages@pytorch.org",
"download_url": null,
"platform": null,
"description": "# FBGEMM_GPU\n\n[![FBGEMM_GPU-CPU CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cpu.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cpu.yml)\n[![FBGEMM_GPU-CUDA CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cuda.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_cuda.yml)\n[![FBGEMM_GPU-ROCm CI](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_rocm.yml/badge.svg)](https://github.com/pytorch/FBGEMM/actions/workflows/fbgemm_gpu_ci_rocm.yml)\n\nFBGEMM_GPU (FBGEMM GPU Kernels Library) is a collection of high-performance\nPyTorch GPU operator libraries for training and inference. The library provides\nefficient table batched embedding bag, data layout transformation, and\nquantization supports.\n\nFBGEMM_GPU is currently tested with CUDA 12.1 and 11.8 in CI, and with PyTorch\npackages (2.1+) that are built against those CUDA versions.\n\nSee the full [Documentation](https://pytorch.org/FBGEMM) for more information\non building, installing, and developing with FBGEMM_GPU, as well as the most\nup-to-date support matrix for this library.\n\n\n## Join the FBGEMM_GPU Community\n\nFor questions, support, news updates, or feature requests, please feel free to:\n\n* File a ticket in [GitHub Issues](https://github.com/pytorch/FBGEMM/issues)\n* Post a discussion in [GitHub Discussions](https://github.com/pytorch/FBGEMM/discussions)\n* Reach out to us on the `#fbgemm` channel in [PyTorch Slack](https://bit.ly/ptslack)\n\nFor contributions, please see the [`CONTRIBUTING`](../CONTRIBUTING.md) file for\nways to help out.\n\n\n## License\n\nFBGEMM_GPU is BSD licensed, as found in the [`LICENSE`](../LICENSE) file.\n",
"bugtrack_url": null,
"license": "BSD-3",
"summary": null,
"version": "2024.9.16",
"project_urls": {
"Homepage": "https://github.com/pytorch/fbgemm"
},
"split_keywords": [
"pytorch",
" recommendation models",
" high performance computing",
" gpu",
" cuda"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "3436c99d9b24065606b265fbba1964e2dfd4a8859f583c21122a1b5f5e2dfc76",
"md5": "4e8dca686b66e8c5acfa8e9866e7cfdd",
"sha256": "259673c856b3a9ff5b2760323f68c04557e23011cde8ff965ab7ea4eeb265d9a"
},
"downloads": -1,
"filename": "fbgemm_gpu_nightly_genai-2024.9.16-cp310-cp310-manylinux2014_x86_64.whl",
"has_sig": false,
"md5_digest": "4e8dca686b66e8c5acfa8e9866e7cfdd",
"packagetype": "bdist_wheel",
"python_version": "cp310",
"requires_python": null,
"size": 3347826,
"upload_time": "2024-09-16T14:02:12",
"upload_time_iso_8601": "2024-09-16T14:02:12.041140Z",
"url": "https://files.pythonhosted.org/packages/34/36/c99d9b24065606b265fbba1964e2dfd4a8859f583c21122a1b5f5e2dfc76/fbgemm_gpu_nightly_genai-2024.9.16-cp310-cp310-manylinux2014_x86_64.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "b6dae1c4a34fd324f659c9a29c9a647bebb4c132e7aaeaffff4258469411c64a",
"md5": "dffe3c66673a004d5e955f3fda9983ce",
"sha256": "04b739d688661e6eb6d71808ac18a75bcc604470acc8a0f399f5a730337fc1d5"
},
"downloads": -1,
"filename": "fbgemm_gpu_nightly_genai-2024.9.16-cp311-cp311-manylinux2014_x86_64.whl",
"has_sig": false,
"md5_digest": "dffe3c66673a004d5e955f3fda9983ce",
"packagetype": "bdist_wheel",
"python_version": "cp311",
"requires_python": null,
"size": 3347832,
"upload_time": "2024-09-16T14:02:42",
"upload_time_iso_8601": "2024-09-16T14:02:42.602380Z",
"url": "https://files.pythonhosted.org/packages/b6/da/e1c4a34fd324f659c9a29c9a647bebb4c132e7aaeaffff4258469411c64a/fbgemm_gpu_nightly_genai-2024.9.16-cp311-cp311-manylinux2014_x86_64.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "4035bdfd266a5c362dd4e7405e96e5722268c6e3d631be56f335fc09a61e91a1",
"md5": "a4404b2702cbdd7f3470c8a303135eef",
"sha256": "85ce307d36fa4c543432b817afc9429fe7a0fbd7b0a91a34775010db06f81120"
},
"downloads": -1,
"filename": "fbgemm_gpu_nightly_genai-2024.9.16-cp312-cp312-manylinux2014_x86_64.whl",
"has_sig": false,
"md5_digest": "a4404b2702cbdd7f3470c8a303135eef",
"packagetype": "bdist_wheel",
"python_version": "cp312",
"requires_python": null,
"size": 3347832,
"upload_time": "2024-09-16T14:06:09",
"upload_time_iso_8601": "2024-09-16T14:06:09.191920Z",
"url": "https://files.pythonhosted.org/packages/40/35/bdfd266a5c362dd4e7405e96e5722268c6e3d631be56f335fc09a61e91a1/fbgemm_gpu_nightly_genai-2024.9.16-cp312-cp312-manylinux2014_x86_64.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "cf340fcc5c418c902ca0e30e67c4bb5445c3c61c7509299b5e8726c59c1fcc4e",
"md5": "1f076102635879bc54bf361f993c9c9c",
"sha256": "fd91df948d03058afb33be9873894e9539257b4769a824c34aa9a1a013c4da9e"
},
"downloads": -1,
"filename": "fbgemm_gpu_nightly_genai-2024.9.16-cp39-cp39-manylinux2014_x86_64.whl",
"has_sig": false,
"md5_digest": "1f076102635879bc54bf361f993c9c9c",
"packagetype": "bdist_wheel",
"python_version": "cp39",
"requires_python": null,
"size": 3457958,
"upload_time": "2024-09-16T14:08:02",
"upload_time_iso_8601": "2024-09-16T14:08:02.779893Z",
"url": "https://files.pythonhosted.org/packages/cf/34/0fcc5c418c902ca0e30e67c4bb5445c3c61c7509299b5e8726c59c1fcc4e/fbgemm_gpu_nightly_genai-2024.9.16-cp39-cp39-manylinux2014_x86_64.whl",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-09-16 14:02:12",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "pytorch",
"github_project": "fbgemm",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "fbgemm-gpu-nightly-genai"
}