We’re evaluating whether AI copilot-style workflow generation actually changes the game for us. The pitch is straightforward: describe what you want in plain English, the AI generates the workflow, and you’re done. No months of back-and-forth with developers, no learning another visual builder language.
But I’m skeptical because I’ve seen this movie before. I’ve used AI tools that claim to turn descriptions into code, and the output is usually 60% there. You spend half the time fixing what the AI generated as you would have spent building it from scratch.
What I really want to know is whether anyone has actually deployed workflows that came directly from AI generation without significant rework. Not just one-off test cases, but real production automation that your team is actually using and maintaining.
I’m trying to understand if we should budget for a 30% rework factor, a 50% factor, or if some of you are actually getting usable workflows on the first pass. And more importantly, when you do get rework needed, what’s the typical issue? Is it logic gaps? Integration failures? Something else?
How realistic is it to actually skip the coding step, or are we fooling ourselves?