Raw data
{
"_id": null,
"home_page": "https://github.com/lucidrains/mixture-of-attention",
"name": "mixture-of-attention",
"maintainer": "",
"docs_url": null,
"requires_python": "",
"maintainer_email": "",
"keywords": "artificial intelligence,deep learning,transformers,attention mechanism,mixture-of-experts,routed attention",
"author": "Phil Wang",
"author_email": "lucidrains@gmail.com",
"download_url": "https://files.pythonhosted.org/packages/05/5a/45a8523ba07e86df09136480b1c60599209192f57bfa270ad00867d26aee/mixture-of-attention-0.0.24.tar.gz",
"platform": null,
"description": "",
"bugtrack_url": null,
"license": "MIT",
"summary": "Mixture of Attention",
"version": "0.0.24",
"project_urls": {
"Homepage": "https://github.com/lucidrains/mixture-of-attention"
},
"split_keywords": [
"artificial intelligence",
"deep learning",
"transformers",
"attention mechanism",
"mixture-of-experts",
"routed attention"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "9aa963d85c51f4dd1671e472553be877309b7950320544f231a096e504d11752",
"md5": "01ca59a39681614a1d294b210ed2594b",
"sha256": "efa65f9e2853fc19ff623dd3e90e5225d0c5c06e68a731644e44c42b3665e40b"
},
"downloads": -1,
"filename": "mixture_of_attention-0.0.24-py3-none-any.whl",
"has_sig": false,
"md5_digest": "01ca59a39681614a1d294b210ed2594b",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": null,
"size": 10525,
"upload_time": "2023-06-23T22:54:48",
"upload_time_iso_8601": "2023-06-23T22:54:48.376434Z",
"url": "https://files.pythonhosted.org/packages/9a/a9/63d85c51f4dd1671e472553be877309b7950320544f231a096e504d11752/mixture_of_attention-0.0.24-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "055a45a8523ba07e86df09136480b1c60599209192f57bfa270ad00867d26aee",
"md5": "f9ae8fe9f83b40739fd0f3f0d558ba63",
"sha256": "190c5f03d2c51fdcccfdf682025014aac27657735ccef2c2735dfd502dfbfc54"
},
"downloads": -1,
"filename": "mixture-of-attention-0.0.24.tar.gz",
"has_sig": false,
"md5_digest": "f9ae8fe9f83b40739fd0f3f0d558ba63",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 10667,
"upload_time": "2023-06-23T22:54:49",
"upload_time_iso_8601": "2023-06-23T22:54:49.322391Z",
"url": "https://files.pythonhosted.org/packages/05/5a/45a8523ba07e86df09136480b1c60599209192f57bfa270ad00867d26aee/mixture-of-attention-0.0.24.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-06-23 22:54:49",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "lucidrains",
"github_project": "mixture-of-attention",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "mixture-of-attention"
}