Residual Kernel Policy Network (ResKPN): Enhancing Stability and Robustness in RKHS-Based Reinforcement Learning
This repository contains the implementation of the algorithms proposed in our paper: Residual Kernel Policy Network: Enhancing Stability and Robustness in RKHS-Based Reinforcement Learning.
The Residual Kernel Policy Network (ResKPN) is a novel framework designed to improve the stability and robustness of reinforcement learning methods in Reproducing Kernel Hilbert Spaces (RKHS). This repository provides the codebase for reproducing the experiments and comparing ResKPN with other algorithms.
This project builds upon the PureJaxRL framework. To set up the environment, use the requirements.txt file:
pip install -r requirements.txtJAX's ability to leverage accelerators (e.g., GPU/TPU) is crucial for efficient parallel training of environments. For detailed installation instructions and configurations, refer to the Jax installation.
To run an algorithm, simply execute the corresponding Python script. For example, to train using the ResKPN algorithm, use:
python ResKPN.pyTo record the training process, set config["DEBUG"] = True. Training logs will be saved in the logs directory and can be visualized using TensorBoard.
To visualize the training process, run
tensorboard --logdir logsHere is an example visualization comparing PPO and ResKPN in the Walker2D environment:
