3. A combination of LEXPOL with the previous natural language based state embedding algorithm, giving a joint method combing state and action factorization
3. A combination of LEXPOL with the previous natural language based state embedding algorithm, giving a joint method combing state and action factorization
1. Qualitative analysis of LEXPOL (end-to-end learning) and frozen pre-trained single-task policies. We note that LEXPOL successfully disentangles the tasks into fundamental skills, and learns to combine them without a decomposition to primitive actions.
1. Qualitative analysis of LEXPOL (end-to-end learning) and frozen pre-trained single-task policies. We note that LEXPOL successfully disentangles the tasks into fundamental skills, and learns to combine them without a decomposition to primitive actions.
- Learning in both supervised and reinforcement learning contexts.
- Learning in both supervised and reinforcement learning contexts.
- Hierarchical Embeddings: We show that it is possible to break down hierarchical value functions into its core elements by leveraging higher-order decomposition methods in Mathematics like Tucker Decompositions.
- Zero-shot generalization: H-UVFAs can extrapolate to new goals!
- Hierarchical Embeddings: We show that it is possible to break down hierarchical value functions into its core elements by leveraging higher-order decomposition methods in Mathematics like Tucker Decompositions.
- Zero-shot generalization: H-UVFAs can extrapolate to new goals!