/decision-mamba

Decision Mamba: Reinforcement Learning via Sequence Modeling with Selective State Spaces

Primary LanguagePythonMIT LicenseMIT

Decision Mamba

Reinforcement Learning via Sequence Modeling with Selective State Spaces
https://arxiv.org/abs/2403.19925

Architecture

Below is the overview of our main module, the Mamba layer: mambablock

We adopt the basic Transformer-type neural network architecture for the Mamba layer, namely it consists of the token-mixing block and the channel-mixing block. The right-hand side of the figure illustrates the series of operations performed inside the Mamba block. $\sigma$ is the $\mathrm{SiLU}$ activation function and $\odot$ denotes the element-wise product. For more details, see Section 3 of the paper.

Instructions

We provide the corresponding code in two sub-directories: atari containing code for Atari experiments and gym containing code for OpenAI Gym experiments. See corresponding READMEs in each folder for instructions. Scripts should be run from the respective directories.

Acknowledgements

Our code largely relies on the implementations of decision-transformer and mamba. We thank their excellent works.

Citation

If you use our code, or otherwise found our work useful, please cite the accompanying paper:

@article{ota2024decision,
    title   = {Decision Mamba: Reinforcement Learning via Sequence Modeling with Selective State Spaces},
    author  = {Toshihiro Ota},
    journal = {arXiv preprint arXiv:2403.19925},
    year    = {2024}
}