-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
Sequence Modeling RL
SOTA RL sequence modelers use transformers and attention which enable long range dependencies and parallel inference, however they have quadratic complexity in sequence length and hidden space dimensionality (lecture 13 below).
- CS885 Lecture 13
- Mamba: Linear-Time Sequence Modeling with Selective State Spaces
- Deep Transformer Q-Networks for Partially Observable Reinforcement Learning
- Decision Transformer: Reinforcement Learning via Sequence Modeling
- Efficiently Modeling Long Sequences with Structured State Spaces
- HiPPO: Recurrent Memory with Optimal Polynomial Projections
Partially Observable RL, DRQN
Constrained RL
Constrained reinforcement learning ensures safety in RL, it helps to incorporate constrained objectives (lecture 10 below).
Distributional RL
Diabetes RL
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels