Document Type
Conference Proceeding
Publication Title
Proceedings of Machine Learning Research
Abstract
Perceived signals in real-world scenarios are usually high-dimensional and noisy, and finding and using their representation that contains essential and sufficient information required by downstream decision-making tasks will help improve computational efficiency and generalization ability in the tasks. In this paper, we focus on partially observable environments and propose to learn a minimal set of state representations that capture sufficient information for decision-making, termed Action-Sufficient state Representations (ASRs). We build a generative environment model for the structural relationships among variables in the system and present a principled way to characterize ASRs based on structural constraints and the goal of maximizing cumulative reward in policy learning. We then develop a structured sequential Variational Auto-Encoder to estimate the environment model and extract ASRs. Our empirical results on CarRacing and VizDoom demonstrate a clear advantage of learning and using ASRs for policy learning. Moreover, the estimated environment model and ASRs allow learning behaviors from imagined outcomes in the compact latent space to improve sample efficiency.
First Page
9260
Last Page
9279
Publication Date
7-2022
Keywords
Computational efficiency, Learning systems, Machine learning
Recommended Citation
B. Huang, et al, "Action-Sufficient State Representation Learning for Control with Structural Constraints", in 39th Intl. Conf. on Machine Learning (ICML 2022), PMLR, vol 162, pp. 9260-9279, 2022. Available: https://proceedings.mlr.press/v162/huang22f/huang22f.pdf
Comments
IR conditions: non-described
Open Access version available on PMLR