academia-rl


Nameacademia-rl JSON
Version 0.2.0 PyPI version JSON
download
home_pageNone
SummaryEasy-to-use tools for Curriculum Learning
upload_time2024-10-01 18:20:38
maintainerNone
docs_urlNone
authorNone
requires_python>=3.9
licenseNone
keywords curriculum-learning reinforcement-learning
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage
            # academia

This package’s purpose is to provide easy-to-use tools for Curriculum Learning.
It is a part of an engineering thesis at Warsaw University of Technology that 
touches on the topic of curriculum learning.

## Documentation

https://academia.readthedocs.io/

## Experiments

As a part of our engineering thesis, numerous experiments on Curriculum Learning 
have been conducted using this package. Their code and results can be found in 
a [separate repo](https://github.com/krezelj/thesis-experiments).

## Sources
An unordered list of interesting books and papers

### Books

1.  [Reinforcement Learning: An Introduction (Barto, Sutton)](https://web.stanford.edu/class/psych209/Readings/SuttonBartoIPRLBook2ndEd.pdf)
  
### Papers

| Paper Link | Short Description |Related Papers |
|------------|-------------------| - |
| [Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey](https://arxiv.org/pdf/2003.04960.pdf?fbclid=IwAR3n0MndHpbiWI1-Wfds5jTXSkwXwpo1mf7jaK-64J4heyYOnYO76qnEWCE) | Survey of curriculum learning papers. Formalises curriuclum learning based on a variety of attributes and gives a good introduction into the topic. ||
| [Curriculum Design for Machine Learners in Sequential Decision Tasks](https://beipeng.github.io/files/2018ieee-tetci-peng.pdf) | Talks about curricula desgined by non-experts (i.e. people who do not know much/anything about a given domain). Uses "dog training" game as the basis of their experiments. Users can design a curriculum by sequencing any number of tasks in more or less complex environments. A target (more difficult) task is also provided to the user but they are not allowed to include it in the curriculum. A trainer model is used to go through the curriculum and provide feedback to the agent. They measure how good a curriculum is by the number of feedbacks that the trainer has to give to the agent i.e. if a curriculum is well designed the agent will require a relatively smaller number of feedbacks from the trainer to move on to the next task. They use three different trainer behaviours and show that the type of the trainer oes not influence the impact of curriuclum design i.e. if a curriculum is well designed under one trainer it is also well designed under another trainer. Another condition for a curriculum to be considered good is that the number of feedbacks over the entire curriculum (with the target task) should be smaller than the number of feedbacks when training on the target task alone. Results show that non-experts can design a better-than-random curriculum when it comes to reducing number of feedbacks on the target task alone but are not better-than-random in desigining a curriculum that decreases the overall number of feedbacks. | [Language and Policy Learning from Human-delivered Feedback](https://beipeng.github.io/files/2015icra-peng.pdf),<br/><br/> [Learning behaviors via human-delivered discrete feedback](https://link.springer.com/article/10.1007/s10458-015-9283-7)
| [Proximal Policy Optimization Algorithms](https://arxiv.org/pdf/1707.06347.pdf) | Not directly related to Curriculum Learning but related to Reinforcement Learning. | 
|[A Deep Hierarchical Approach to Lifelong Learning in Minecraft](https://arxiv.org/pdf/1604.07255.pdf)|Haven't read it yet, not directly connected to CL but should still be helpful||

            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "academia-rl",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.9",
    "maintainer_email": null,
    "keywords": "curriculum-learning, reinforcement-learning",
    "author": null,
    "author_email": "krezelj <jasiekk9@gmail.com>, guts <szymon.gut01@gmail.com>, maciejors <maciej.orslowski@gmail.com>",
    "download_url": "https://files.pythonhosted.org/packages/10/7c/b69e89b413cfc8c8373a85a231837dfb13b96e3c7bf298d21c529597ebf5/academia_rl-0.2.0.tar.gz",
    "platform": null,
    "description": "# academia\n\nThis package\u2019s purpose is to provide easy-to-use tools for Curriculum Learning.\nIt is a part of an engineering thesis at Warsaw University of Technology that \ntouches on the topic of curriculum learning.\n\n## Documentation\n\nhttps://academia.readthedocs.io/\n\n## Experiments\n\nAs a part of our engineering thesis, numerous experiments on Curriculum Learning \nhave been conducted using this package. Their code and results can be found in \na [separate repo](https://github.com/krezelj/thesis-experiments).\n\n## Sources\nAn unordered list of interesting books and papers\n\n### Books\n\n1.  [Reinforcement Learning: An Introduction (Barto, Sutton)](https://web.stanford.edu/class/psych209/Readings/SuttonBartoIPRLBook2ndEd.pdf)\n  \n### Papers\n\n| Paper Link | Short Description |Related Papers |\n|------------|-------------------| - |\n| [Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey](https://arxiv.org/pdf/2003.04960.pdf?fbclid=IwAR3n0MndHpbiWI1-Wfds5jTXSkwXwpo1mf7jaK-64J4heyYOnYO76qnEWCE) | Survey of curriculum learning papers. Formalises curriuclum learning based on a variety of attributes and gives a good introduction into the topic. ||\n| [Curriculum Design for Machine Learners in Sequential Decision Tasks](https://beipeng.github.io/files/2018ieee-tetci-peng.pdf) | Talks about curricula desgined by non-experts (i.e. people who do not know much/anything about a given domain). Uses \"dog training\" game as the basis of their experiments. Users can design a curriculum by sequencing any number of tasks in more or less complex environments. A target (more difficult) task is also provided to the user but they are not allowed to include it in the curriculum. A trainer model is used to go through the curriculum and provide feedback to the agent. They measure how good a curriculum is by the number of feedbacks that the trainer has to give to the agent i.e. if a curriculum is well designed the agent will require a relatively smaller number of feedbacks from the trainer to move on to the next task. They use three different trainer behaviours and show that the type of the trainer oes not influence the impact of curriuclum design i.e. if a curriculum is well designed under one trainer it is also well designed under another trainer. Another condition for a curriculum to be considered good is that the number of feedbacks over the entire curriculum (with the target task) should be smaller than the number of feedbacks when training on the target task alone. Results show that non-experts can design a better-than-random curriculum when it comes to reducing number of feedbacks on the target task alone but are not better-than-random in desigining a curriculum that decreases the overall number of feedbacks. | [Language and Policy Learning from Human-delivered Feedback](https://beipeng.github.io/files/2015icra-peng.pdf),<br/><br/> [Learning behaviors via human-delivered discrete feedback](https://link.springer.com/article/10.1007/s10458-015-9283-7)\n| [Proximal Policy Optimization Algorithms](https://arxiv.org/pdf/1707.06347.pdf) | Not directly related to Curriculum Learning but related to Reinforcement Learning. | \n|[A Deep Hierarchical Approach to Lifelong Learning in Minecraft](https://arxiv.org/pdf/1604.07255.pdf)|Haven't read it yet, not directly connected to CL but should still be helpful||\n",
    "bugtrack_url": null,
    "license": null,
    "summary": "Easy-to-use tools for Curriculum Learning",
    "version": "0.2.0",
    "project_urls": {
        "Documentation": "https://academia.readthedocs.io/",
        "Source code": "https://github.com/krezelj/academia"
    },
    "split_keywords": [
        "curriculum-learning",
        " reinforcement-learning"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "73203cbd2fe1672e876157e2e3356680b786d0be90beba9be411fc290e2670e6",
                "md5": "2f8035861f29ed0b11cdd55c49ade7d6",
                "sha256": "41b2d341742abddfa1e8455eead96cb65c94cfc909224db498a5a462bb5cd516"
            },
            "downloads": -1,
            "filename": "academia_rl-0.2.0-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "2f8035861f29ed0b11cdd55c49ade7d6",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.9",
            "size": 76857,
            "upload_time": "2024-10-01T18:20:34",
            "upload_time_iso_8601": "2024-10-01T18:20:34.816769Z",
            "url": "https://files.pythonhosted.org/packages/73/20/3cbd2fe1672e876157e2e3356680b786d0be90beba9be411fc290e2670e6/academia_rl-0.2.0-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "107cb69e89b413cfc8c8373a85a231837dfb13b96e3c7bf298d21c529597ebf5",
                "md5": "a170fb2bbdd3c6940a2d6d0d91de0a0b",
                "sha256": "787a280d65b368633914f3450adcf26d512a1054004d5154a7ea4be14de34ad5"
            },
            "downloads": -1,
            "filename": "academia_rl-0.2.0.tar.gz",
            "has_sig": false,
            "md5_digest": "a170fb2bbdd3c6940a2d6d0d91de0a0b",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.9",
            "size": 53023,
            "upload_time": "2024-10-01T18:20:38",
            "upload_time_iso_8601": "2024-10-01T18:20:38.709531Z",
            "url": "https://files.pythonhosted.org/packages/10/7c/b69e89b413cfc8c8373a85a231837dfb13b96e3c7bf298d21c529597ebf5/academia_rl-0.2.0.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-10-01 18:20:38",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "krezelj",
    "github_project": "academia",
    "travis_ci": false,
    "coveralls": true,
    "github_actions": false,
    "lcname": "academia-rl"
}
        
Elapsed time: 4.43006s