We built an inference engine at Kog generating 10,000 tokens per second per request, where standard providers sit around 100 t/s.
We are looking for teams working on generation-heavy workflows like full-stack Python app generation from a single prompt to validate this together.
What would that concretely change for the Reflex team?