SBC generates virtual environments via state blocking to expose agents to diverse suboptimal partner policies, yielding superior zero-shot coordination performance including with humans.
Responsive safety in reinforcement learning by pid lagrangian methods
2 Pith papers cite this work. Polarity classification is still indexing.
citation-role summary
citation-polarity summary
fields
cs.LG 2years
2026 2verdicts
UNVERDICTED 2roles
background 1polarities
background 1representative citing papers
AdamFLIP treats PDE constraint residuals in PINNs as a controlled dynamical system, computes Lagrange multipliers via feedback linearization to drive residuals to zero, and applies Adam-style adaptation to the resulting gradient for scalable hard-constrained training.
citing papers explorer
-
Shaping Zero-Shot Coordination via State Blocking
SBC generates virtual environments via state blocking to expose agents to diverse suboptimal partner policies, yielding superior zero-shot coordination performance including with humans.
-
AdamFLIP: Adaptive Momentum Feedback Linearization Optimization for Hard Constrained PINN Training
AdamFLIP treats PDE constraint residuals in PINNs as a controlled dynamical system, computes Lagrange multipliers via feedback linearization to drive residuals to zero, and applies Adam-style adaptation to the resulting gradient for scalable hard-constrained training.