💬 Host Commentary

For this Ship It Conversations episode, I wanted to get past the LinkedIn hype cycle around “AI agents for DevOps” and talk to someone who actually wired one into their stack without losing sleep over it.

Gracious has been doing exactly that with TARS, a human-in-the-loop fixer bot that plugs into CI/CD, GitHub, and your containers. What I like about his story is it didn’t start as “let’s build an AIOps platform.” It started with a very boring, very real problem: chasing Docker logs and doing the same incident steps over and over. First he built Friday, a status bot that could safely poke at containers and tell him what died and why. Then he layered TARS on top to correlate with commits, suggest where to look, and eventually help drive rollbacks, all behind hard guardrails.

The guardrail piece is what made this worth recording. He’s aggressively narrow with what each agent can see and do. One workflow can only run safe Docker commands. Another can read GitHub but can’t touch infra. Actions that change the world, like redeploying the last good build, require an explicit phrase from a human, and even then there’s a second layer of validation in the workflow itself. It’s not perfect or “formally verified” or any of that, but it’s a real example of segmenting incident response into sub-workflows and keeping the agent boxed in at each step.

I also appreciated his answer to “where should teams start?” His take: you don’t start with an LLM. You start by writing down your own process, breaking it into steps, and turning those into programs and sub-flows. Only then do you drop an LLM in as glue or a helper. Same thing with skepticism about fully autonomous agents. He’s pro-automation, but he’s still very clear that human judgment, gut checks, and validating every LLM output before acting on it are the non-negotiables.

If you’re the platform / SRE / DevOps person who keeps getting asked about agents, AIOps, or “can we use AI to fix incidents,” this conversation should give you a concrete example and a vocabulary for pushing toward human-in-the-loop systems instead of “give the bot SSH and hope.” Links to TARS, Friday, and Gracious’s posts are down below, along with the other Ship It Conversations episodes.

📝 Show Notes

This is a guest conversation episode of Ship It Weekly (separate from the weekly news recaps).

In this Ship It: Conversations episode I talk with Gracious James Eluvathingal about TARS, his “human-in-the-loop” fixer bot wired into CI/CD.

We get into why he built it in the first place, how he stitches together n8n, GitHub, SSH, and guardrailed commands, and what it actually looks like when an AI agent helps with incident response without being allowed to nuke prod. We also dig into rollback phases, where humans stay in the loop, and why validating every LLM output before acting on it is the single most important guardrail.

If you’re curious about AI agents in pipelines but hate the idea of a fully autonomous “ops bot,” this one is very much about the middle ground: segmenting workflows, limiting blast radius, and using agents to reduce toil instead of replace engineers.

Gracious also walks through where he’d like to take TARS next (Terraform, infra-level decisions, more tools) and gives some solid advice for teams who want to experiment with agents in CI/CD without starting with “let’s give it root and see what happens.”

Links from the episode:

Gracious on LinkedIn: https://www.linkedin.com/in/gracious-james-eluvathingal

TARS overview post: https://www.linkedin.com/posts/gracious-james-eluvathingal_aiagents-devops-automation-activity-7391064503892987904-psQ4

If you found this useful, share it with the person on your team who’s poking at AI automation and worrying about guardrails.

More information on our website: https://shipitweekly.fm