Jingkang Yang*,1 Yuhao Dong*,2,5 Shuai Liu*,3,5 Bo Li*,1
Ziyue Wang†,1 Chencheng Jiang†,4 Haoran Tan†,3 Jiamu Kang†,2
Yuanhan Zhang1 Kaiyang Zhou1 Ziwei Liu1,5,✉
3Beijing University of Posts and Telecommunications
4Xi'an Jiaotong University 5Shanghai AI Laboratory
* Equal Contribution † Equal Engineering Contribution ✉ Corresponding Author
Project Page | Octopus Paper | Demo Video
Octopus is a novel VLM designed to proficiently decipher an agent’s vision and textual task objectives and to formulate intricate action sequences and generate executable code. We provide two models based on the following architectures. Please click
OctoVerse contains three sub-worlds
OS (tested) | Environment Goal | |
---|---|---|
OctoGibson | Ubuntu 20.04 | 500 Tasks on OmniGibson |
OctoGTA | Windows 11 | 20 Tasks to evaluate transfer learning |
OctoMC | Ubuntu/Windows | 20 Tasks to evaluate transfer learning on MineCraft worlds, such as making an axe. |
- Training data collection pipeline in
octogibson
environment - Evaluation pipeline in
octogibson
environment - Evaluation pipeline in
octogta
environment - Training pipeline of the
octopus
model
Contact: Leave issue or contact [email protected]
and [email protected]
. We are on call to respond.
[2023-10]
- 🤗 Introducing Project Octopus' homepage: https://choiszt.github.io/Octopus.
- 🤗 Check our paper introducing Octopus in details.
- Training Data Collection: For data collection from
octogibson
environment, you need to set up two conda environments:omnigibson
andgpt4
. Theomnigibson
environment has an agent to act following the instruction fromgpt4
environment. Please checkout here for detailed information. - Evaluation in OctoGibson: We provide the pipeline that the simulator sends messages to the Octopus server and gets responses to control the agent.
- Evaluation in OctoGTA: We provide instructions, code, and MOD so that the Octopus can complete tasks in the GTA environment. Please checkout here for detailed information.
- Octopus Training: We provide code for training Octopus. Please checkout here for detailed information.
If you found this repository useful, please consider citing:
@article{yang2023octopus,
title = {Octopus: Embodied Vision-Language Programmer from Environmental Feedback},
author = {Jingkang Yang and Yuhao Dong and Shuai Liu and Bo Li and Ziyue Wang and Chencheng Jiang and Haoran Tan and Jiamu Kang and Yuanhan Zhang and Kaiyang Zhou and Ziwei Liu},
year = {2023},
}
We thank the OmniGibson team for their help and great contribution to the open-source community.