All these agent documentations seem to compete for the most complex set of flow charts imaginable without ever mentioning what the Rube Goldberg machine is supposed to accomplish. Given that the real output in open source of these contraptions is zero, it seems that the flow charts are the goal. Some kind of modern art.
It doesn't replace core algorithms. It plumbs things together. It means you're not having to write the framework to connect things, your algos are still going to have the same problems as they had before.
This is just a style I've seen a lot of people who are a generation or so younger than me enjoy.
I'm not expected to write docs the way my father's generation did (thank god), so I don't expect them to write the docs the way I would. If this gets people engaged and excited, I lose nothing, they get something, we're fine.
As to the LLM generation claim, I don't care if it is or it isn't. The project seems legit, they're making claims that 3rd parties have verified ("Community Projects"), it looks useful and interesting, so I might spend more time with it soon.
So it’s some brittle crap built on verl, which is already pretty much train by config (and makes breaking changes with _every single commit_), with no documentation, no examples, and no clear purpose? Heck yeah Microsoft
A framework for optimizing LLM agents, including but not limited to RL. You can even do fine tuning, they have an example with unsloth in there.
The design of this is pretty nice, it's based on a very simple to add instrumentation to your agent and the rest happens in parallel while your workload runs which is awesome.
You can probably do also what DSPy does for optimizing prompts but without having to rewrite using the DSPy API which can be a big win.
reply