Skill Disentanglement in Reproducing Kernel Hilbert Space
Abstract
Unsupervised Skill Discovery aims at learning diverse skills without any extrinsic rewards and leverage them as prior for learning a variety of downstream tasks. Existing approaches to unsupervised reinforcement learning typically involve discovering skills through empowerment-driven techniques or by maximizing entropy to encourage exploration. However, this mutual information objective often results in either static skills that discourage exploration or maximise coverage at the expense of non-discriminable skills. Instead of focusing only on maximizing bounds on f-divergence, we combine it with Integral Probability Metrics to maximize the distance between distributions to promote behavioural diversity and enforce disentanglement. Our method, Hilbert Unsupervised Skill Discovery (HUSD), provides an additional objective that seeks to obtain exploration and separability of state-skill pairs by maximizing the Maximum Mean Discrepancy between the joint distribution of skills and states and the product of their marginals in Reproducing Kernel Hilbert Space. Our results on Unsupervised RL Benchmark show that HUSD outperforms previous exploration algorithms on state-based tasks.
Cite
Text
Dave and Rueckert. "Skill Disentanglement in Reproducing Kernel Hilbert Space." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I15.33774Markdown
[Dave and Rueckert. "Skill Disentanglement in Reproducing Kernel Hilbert Space." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/dave2025aaai-skill/) doi:10.1609/AAAI.V39I15.33774BibTeX
@inproceedings{dave2025aaai-skill,
title = {{Skill Disentanglement in Reproducing Kernel Hilbert Space}},
author = {Dave, Vedant and Rueckert, Elmar},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {16153-16162},
doi = {10.1609/AAAI.V39I15.33774},
url = {https://mlanthology.org/aaai/2025/dave2025aaai-skill/}
}