robocat


Namerobocat JSON
Version 0.1.0 PyPI version JSON
download
home_pagehttps://github.com/kyegomez/RoboCAT
SummaryRobo CAT- Pytorch
upload_time2023-09-04 15:44:13
maintainer
docs_urlNone
authorKye Gomez
requires_python>=3.6,<4.0
licenseMIT
keywords artificial intelligence deep learning transformers attention mechanism robotics
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            [![Multi-Modality](agorabanner.png)](https://discord.gg/qUtxnK2NMf)


# ROBOTCAT
![ROBOCAT MODELS](robocat.png)

RoboCAT is a self-improving foundation agent for robotic manipulation developed by DeepMind Robotics. The model architecture of RoboCAT is similar to the RT-1 model. It follows a tokenization approach where robotics images, proprioception, and future actions are tokenized. 

* Help with the implementation in the Agora discord, ![Discord](https://img.shields.io/discord/999382051935506503)

----

# Install
`pip install robotcat` 

---

## Basic Usage

```python
import torch 
from robocat.model import RoboCat

model = RoboCat()

video = torch.randn(2, 3, 6, 224, 224)
instructions = [
    "bring me that apple"
]

result = model.forward(video, instructions)
print(result)

```


## Generate Data
* RoboCat can self generate data with images and or video, to use simply use the generator classes -> then tokenize and or tensorize using your custom strategy

```python
#for images
from robotcat import ImageDataGenerator

generator = ImageDataGenerator()

prompt = "Robot picking up cup in kitchen"
generator.generate(prompt)
```

and or for videos

```python
from robocat import VideoDataGenerator, RoboCat

generator = VideoDataGenerator()
model = RoboCat()

prompt = "Robot picking up cup in kitchen in first person"
video = generator.generate(prompt)
video = transform_to_tensors(video) #(batch, frames, actions, bins)

model(video, prompt)

```
----


## Architecture
The architecture consists of the following key components:

1. Tokenizer: RoboCAT learns a tokenizer for robotics images. It tokenizes proprioception and future actions in a straightforward manner, enabling the prediction of future action tokens.

2. Transformer Model: The tokenized inputs are fed into a Transformer model. The Transformer model predicts future action sequences based on the input tokens. This allows the model to perform a wide range of robotic tasks using a unified interface.

3. Action Spaces: RoboCAT predicts Cartesian 4 or 6 degrees of freedom (DoF) cartesian velocities for the arm and 1 DoF (parallel jaw gripper) or 8 DoF (3-finger) for the hand. This flexible approach enables the model to handle action spaces of different sizes and variable proprioception sizes.

The architecture of RoboCAT allows for the integration of multiple robot embodiments with a unified interface. By predicting the appropriate number of tokens based on the robot's morphology, the model can effectively scale without the need for separate prediction heads for each embodiment.


## Misc Components

### Generalization and Transfer Learning
The RoboCAT paper focuses on studying generalization and transfer learning. It explores how training on one domain benefits testing on another and investigates the effectiveness of transfer learning from simulation to the real world. The authors provide empirical data on cross-task transfer, architecture scaling, and tokenization strategies for perception.

### Evaluation and Automated Testing
RoboCAT emphasizes the importance of rigorous evaluation and presents methodologies for automated evaluation of multi-task policies in real-world settings. The paper provides details on evaluation protocols, data collection, and comparative analysis of different models and approaches.

### Real-World Robotic Tasks
The paper highlights the challenges of real-world robotics tasks and the significance of cross-robot transfer. The authors showcase consistent results across multiple robots and action spaces, demonstrating the value of collecting real-world data for training and evaluation. The effort put into data set detailing and evaluation protocols is commendable.

### Future Directions and Reproducibility
The authors acknowledge the ongoing challenge of reproducibility in robotics research. They emphasize the need for independent replication in different labs and variations in manipulation tasks and hardware. The paper raises questions about the impact of experimental choices and engineering decisions on research outcomes and

 calls for advancements in evaluation methodologies.

## Conclusion

The RoboCAT paper presents a self-improving foundation agent for robotic manipulation that addresses the challenges of generalization and transfer learning in the field of robotics. It offers insights into the model architecture, requirements, and experimental findings. The extensive empirical data, evaluation protocols, and comparisons provide valuable contributions to the research community.

# Roadmap

* Functional prototype

* Integrate VQGAN to generate an image when it has not encountered an known environment

` environment observation -> environment familarity rating [0.0-1.0] -> generate data if lower then [0.5] -> finetune -> action`

* Train on massive datasets

* Finetune as specified on paper

* Release as paid API

* Integrate more modalities like hearing, 3d mapping, nerfs, videos, lidar, locomotion, and the whole lot!


## Citations

```bibtex
@article{Bousmalis2023RoboCat,
    title   = {RoboCat: A Self-Improving Foundation Agent for Robotic Manipulation},
    author  = {Konstantinos Bousmalis*, Giulia Vezzani*, Dushyant Rao*, Coline Devin*, Alex X. Lee*, Maria Bauza*, Todor Davchev*, Yuxiang Zhou*, Agrim Gupta*,1, Akhil Raju, Antoine Laurens, Claudio Fantacci, Valentin Dalibard, Martina Zambelli, Murilo Martins, Rugile Pevceviciute, Michiel Blokzijl, Misha Denil, Nathan Batchelor, Thomas Lampe, Emilio Parisotto, Konrad Żołna, Scott Reed, Sergio Gómez Colmenarejo, Jon Scholz, Abbas Abdolmaleki, Oliver Groth, Jean-Baptiste Regli, Oleg Sushkov, Tom Rothörl, José Enrique Chen, Yusuf Aytar, Dave Barker, Joy Ortiz, Martin Riedmiller, Jost Tobias Springenberg, Raia Hadsell†, Francesco Nori† and Nicolas Heess},
    journal = {ArXiv},
    year    = {2023}
}
```




            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/kyegomez/RoboCAT",
    "name": "robocat",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.6,<4.0",
    "maintainer_email": "",
    "keywords": "artificial intelligence,deep learning,transformers,attention mechanism,robotics",
    "author": "Kye Gomez",
    "author_email": "kye@apac.ai",
    "download_url": "https://files.pythonhosted.org/packages/37/0f/27a6039c190236612ffaf7a6f0e2d0350c566c0417dce3d1a4320c9ebe96/robocat-0.1.0.tar.gz",
    "platform": null,
    "description": "[![Multi-Modality](agorabanner.png)](https://discord.gg/qUtxnK2NMf)\n\n\n# ROBOTCAT\n![ROBOCAT MODELS](robocat.png)\n\nRoboCAT is a self-improving foundation agent for robotic manipulation developed by DeepMind Robotics. The model architecture of RoboCAT is similar to the RT-1 model. It follows a tokenization approach where robotics images, proprioception, and future actions are tokenized. \n\n* Help with the implementation in the Agora discord, ![Discord](https://img.shields.io/discord/999382051935506503)\n\n----\n\n# Install\n`pip install robotcat`\u00a0\n\n---\n\n## Basic Usage\n\n```python\nimport torch \nfrom robocat.model import RoboCat\n\nmodel = RoboCat()\n\nvideo = torch.randn(2, 3, 6, 224, 224)\ninstructions = [\n    \"bring me that apple\"\n]\n\nresult = model.forward(video, instructions)\nprint(result)\n\n```\n\n\n## Generate Data\n* RoboCat can self generate data with images and or video, to use simply use the generator classes -> then tokenize and or tensorize using your custom strategy\n\n```python\n#for images\nfrom robotcat import ImageDataGenerator\n\ngenerator = ImageDataGenerator()\n\nprompt = \"Robot picking up cup in kitchen\"\ngenerator.generate(prompt)\n```\n\nand or for videos\n\n```python\nfrom robocat import VideoDataGenerator, RoboCat\n\ngenerator = VideoDataGenerator()\nmodel = RoboCat()\n\nprompt = \"Robot picking up cup in kitchen in first person\"\nvideo = generator.generate(prompt)\nvideo = transform_to_tensors(video) #(batch, frames, actions, bins)\n\nmodel(video, prompt)\n\n```\n----\n\n\n## Architecture\nThe architecture consists of the following key components:\n\n1. Tokenizer: RoboCAT learns a tokenizer for robotics images. It tokenizes proprioception and future actions in a straightforward manner, enabling the prediction of future action tokens.\n\n2. Transformer Model: The tokenized inputs are fed into a Transformer model. The Transformer model predicts future action sequences based on the input tokens. This allows the model to perform a wide range of robotic tasks using a unified interface.\n\n3. Action Spaces: RoboCAT predicts Cartesian 4 or 6 degrees of freedom (DoF) cartesian velocities for the arm and 1 DoF (parallel jaw gripper) or 8 DoF (3-finger) for the hand. This flexible approach enables the model to handle action spaces of different sizes and variable proprioception sizes.\n\nThe architecture of RoboCAT allows for the integration of multiple robot embodiments with a unified interface. By predicting the appropriate number of tokens based on the robot's morphology, the model can effectively scale without the need for separate prediction heads for each embodiment.\n\n\n## Misc Components\n\n### Generalization and Transfer Learning\nThe RoboCAT paper focuses on studying generalization and transfer learning. It explores how training on one domain benefits testing on another and investigates the effectiveness of transfer learning from simulation to the real world. The authors provide empirical data on cross-task transfer, architecture scaling, and tokenization strategies for perception.\n\n### Evaluation and Automated Testing\nRoboCAT emphasizes the importance of rigorous evaluation and presents methodologies for automated evaluation of multi-task policies in real-world settings. The paper provides details on evaluation protocols, data collection, and comparative analysis of different models and approaches.\n\n### Real-World Robotic Tasks\nThe paper highlights the challenges of real-world robotics tasks and the significance of cross-robot transfer. The authors showcase consistent results across multiple robots and action spaces, demonstrating the value of collecting real-world data for training and evaluation. The effort put into data set detailing and evaluation protocols is commendable.\n\n### Future Directions and Reproducibility\nThe authors acknowledge the ongoing challenge of reproducibility in robotics research. They emphasize the need for independent replication in different labs and variations in manipulation tasks and hardware. The paper raises questions about the impact of experimental choices and engineering decisions on research outcomes and\n\n calls for advancements in evaluation methodologies.\n\n## Conclusion\n\nThe RoboCAT paper presents a self-improving foundation agent for robotic manipulation that addresses the challenges of generalization and transfer learning in the field of robotics. It offers insights into the model architecture, requirements, and experimental findings. The extensive empirical data, evaluation protocols, and comparisons provide valuable contributions to the research community.\n\n# Roadmap\n\n* Functional prototype\n\n* Integrate VQGAN to generate an image when it has not encountered an known environment\n\n` environment observation -> environment familarity rating [0.0-1.0] -> generate data if lower then [0.5] -> finetune -> action`\n\n* Train on massive datasets\n\n* Finetune as specified on paper\n\n* Release as paid API\n\n* Integrate more modalities like hearing, 3d mapping, nerfs, videos, lidar, locomotion, and the whole lot!\n\n\n## Citations\n\n```bibtex\n@article{Bousmalis2023RoboCat,\n    title   = {RoboCat: A Self-Improving Foundation Agent for Robotic Manipulation},\n    author  = {Konstantinos Bousmalis*, Giulia Vezzani*, Dushyant Rao*, Coline Devin*, Alex X. Lee*, Maria Bauza*, Todor Davchev*, Yuxiang Zhou*, Agrim Gupta*,1, Akhil Raju, Antoine Laurens, Claudio Fantacci, Valentin Dalibard, Martina Zambelli, Murilo Martins, Rugile Pevceviciute, Michiel Blokzijl, Misha Denil, Nathan Batchelor, Thomas Lampe, Emilio Parisotto, Konrad \u017bo\u0142na, Scott Reed, Sergio G\u00f3mez Colmenarejo, Jon Scholz, Abbas Abdolmaleki, Oliver Groth, Jean-Baptiste Regli, Oleg Sushkov, Tom Roth\u00f6rl, Jos\u00e9 Enrique Chen, Yusuf Aytar, Dave Barker, Joy Ortiz, Martin Riedmiller, Jost Tobias Springenberg, Raia Hadsell\u2020, Francesco Nori\u2020 and Nicolas Heess},\n    journal = {ArXiv},\n    year    = {2023}\n}\n```\n\n\n\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "Robo CAT- Pytorch",
    "version": "0.1.0",
    "project_urls": {
        "Documentation": "https://github.com/kyegomez/RoboCAT",
        "Homepage": "https://github.com/kyegomez/RoboCAT",
        "Repository": "https://github.com/kyegomez/RoboCAT"
    },
    "split_keywords": [
        "artificial intelligence",
        "deep learning",
        "transformers",
        "attention mechanism",
        "robotics"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "05917743a8470fd66f518e9f9e1d9486ed8ce5a99789e9ee201f332c49814085",
                "md5": "1244cc7e401e44d6ba9bedb8b682a3e1",
                "sha256": "fbb902dd7e09efbbf639f2c78406cd0aff3ed8e18a2d5b380fff4e138a9b7fb5"
            },
            "downloads": -1,
            "filename": "robocat-0.1.0-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "1244cc7e401e44d6ba9bedb8b682a3e1",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.6,<4.0",
            "size": 25721,
            "upload_time": "2023-09-04T15:44:11",
            "upload_time_iso_8601": "2023-09-04T15:44:11.889070Z",
            "url": "https://files.pythonhosted.org/packages/05/91/7743a8470fd66f518e9f9e1d9486ed8ce5a99789e9ee201f332c49814085/robocat-0.1.0-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "370f27a6039c190236612ffaf7a6f0e2d0350c566c0417dce3d1a4320c9ebe96",
                "md5": "6eb28d782419cc06485d22099f863bf2",
                "sha256": "0b1ca55a4d507ba84b3ea1e455243176e69f31d024a45482c3c3e5bad3485ab4"
            },
            "downloads": -1,
            "filename": "robocat-0.1.0.tar.gz",
            "has_sig": false,
            "md5_digest": "6eb28d782419cc06485d22099f863bf2",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.6,<4.0",
            "size": 26761,
            "upload_time": "2023-09-04T15:44:13",
            "upload_time_iso_8601": "2023-09-04T15:44:13.475463Z",
            "url": "https://files.pythonhosted.org/packages/37/0f/27a6039c190236612ffaf7a6f0e2d0350c566c0417dce3d1a4320c9ebe96/robocat-0.1.0.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2023-09-04 15:44:13",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "kyegomez",
    "github_project": "RoboCAT",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "robocat"
}
        
Elapsed time: 0.13833s