Raw data
{
"_id": null,
"home_page": "",
"name": "warp-attention",
"maintainer": "",
"docs_url": null,
"requires_python": "",
"maintainer_email": "",
"keywords": "transformers,attention,scaled dot product attention,pytorch",
"author": "demoriarty",
"author_email": "",
"download_url": "https://files.pythonhosted.org/packages/dd/0c/657c25b612250bfb2a032a138157b97819a543142e445c394d42f7049c13/warp_attention-0.1.9.tar.gz",
"platform": null,
"description": "",
"bugtrack_url": null,
"license": "",
"summary": "Warp attention: hardware efficient implementation of scaled dot product attention.",
"version": "0.1.9",
"project_urls": null,
"split_keywords": [
"transformers",
"attention",
"scaled dot product attention",
"pytorch"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "dd0c657c25b612250bfb2a032a138157b97819a543142e445c394d42f7049c13",
"md5": "bd2b0ac1265fe17c8ea31559eb4ec4a9",
"sha256": "5e661cfea1c5b962b3b5d268e76e564a1fe21ceab785a1a175ee493d2e93db03"
},
"downloads": -1,
"filename": "warp_attention-0.1.9.tar.gz",
"has_sig": false,
"md5_digest": "bd2b0ac1265fe17c8ea31559eb4ec4a9",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 83223084,
"upload_time": "2023-10-07T13:53:07",
"upload_time_iso_8601": "2023-10-07T13:53:07.551700Z",
"url": "https://files.pythonhosted.org/packages/dd/0c/657c25b612250bfb2a032a138157b97819a543142e445c394d42f7049c13/warp_attention-0.1.9.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-10-07 13:53:07",
"github": false,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"lcname": "warp-attention"
}