-
bomberland
Bomberland: a multi-agent AI competition based on Bomberman. This repository contains both starter / hello world kits + the engine source code
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
I made a similar game a few years ago while in undergrad: https://github.com/aybabtme/bomberman
Nowhere as polished or ambitious, but it just came to mind. Bomberman is a fun game!
Thanks for the feedback! We're working on improving the onboarding flow. Sorry about the Docker link issue - it should link you to a copy of the environment binary so that you can play around without Docker (no docs for this workflow just yet unfortunately).
It's essentially a Bomberman-inspired game, where you program the agents to play in it and can play against other users' agents. You can try it out without an account by cloning one of the starter kits here: https://github.com/CoderOneHQ/bomberland and following the usage instructions (but you'll need to create an account to use the visualizer and to submit agents).
We recommend the Docker flow, but if you get stuck feel free to reach out to me (Joy) or @thegalah (Matt) on our Discord: https://discord.gg/NkfgvRN
No others (at the moment). We're a small team so Bomberland is our current focus - we want to improve the tooling first so that it's easy for people to dive into ML before we introduce other environments.
We do have a mini-project called Ultimate Volleyball (https://github.com/CoderOneHQ/ultimate-volleyball) built on Unity ML-Agents. It's intended more as an introduction to deep reinforcement learning, and we wrote some tutorials for it here if anyone's interested: https://www.gocoder.one/blog/hands-on-introduction-to-deep-r...
https://github.com/DevJac/solve_the_spire
I stretched the truth a bit, I'm actually doing something like "hierarchical model-free reinforcement learning", even so, figuring out how to break the game down to create a hierarchy of agents is a lot of work. Basically, the AI is composed of about 8 different traditional RL agents (neural networks), each deciding a different thing. One chooses which cards to draft, one chooses which actions to take in combat, one chooses which path to take on the map, etc.
It shows definite signs of improvement, but has only reached a point where it can beat the act 1 boss about 50% of the time. I think that is its limit right now. I'm doing policy gradient which is very sample inefficient. I'm going to implement soft-actor-critic and see if it can do better with better sample efficiency.
And to reiterate my original point, I think each developer only has one or two reverse engineering attempts in them. I might otherwise be interested in this AI competition, but reverse engineering the environment to create my own model is just too daunting, I'm so burned out on it already.
Related posts
-
Doubt about numpy's eigen calculation
-
ACTorch: a PyTorch-based deep reinforcement learning framework for fast prototyping
-
Using AI to infer depth information from images in Godot 4 .NET 6 using the MiDaS monocular depth model
-
[D] Maintaining documentation with live results from experiments
-
Volleyball agents trained using competitive self-play [tutorial + project link]