👋 HelloRL: A modular RL framework with a single training function that goes from Actor Critic, to PPO and TD3, making it super easy to swap between them (I just published this today)
|
I learned RL recently, but was unsatisfied with the frameworks available, so a month ago I reached out on here with some ideas and got some great feedback, which has led me to today publishing my library, HelloRL, a modular framework that makes it super easy to go from Actor Critic to TD3. Here is the intro from the repo readme: Why is RL usually so hard? RL algorithms are all similar, but they also have unique implementation details and subtle differences. Every RL framework implements each algorithm from scratch, reproducing many of the same steps across hundreds of lines of code, but with minor implementation differences along the way. Trying to swap between them and keep your code working can be a nightmare. If you want to experiment with a new idea on top of Actor Critic, and then try it on a PPO implementation, you would have to spend hours integrating, and hope you didn’t make a mistake. It’s a minefield — it’s so easy to trip yourself up and get something wrong without realising. Introducing HelloRL HelloRL flips this on its head, with a single HelloRL:
https://github.com/i10e-lab/HelloRL Please leave a star ⭐ if you find it useful. submitted by /u/Illustrious-Egg5459 |