Host Commentary

For this Conversations episode, I wanted to stay anchored on a question that I think is going to matter a lot more over the next couple years.

Not whether AI can help with infrastructure.

Whether it should be trusted anywhere near real infrastructure before it has a place to prove itself first.

That is why this one interested me.

Because Ang Chen is not really pitching “let the agent run prod.” He keeps bringing it back to a safer idea than that. Build a sandbox. Build a digital twin. Let Terraform, CloudFormation, SDK scripts, and even AI-assisted workflows hit that first. Then see what breaks before anything touches the real cloud.

What I liked most is that the conversation did not stay at the vague “AI will change everything” level.

He actually gives a pretty grounded answer for what high fidelity is supposed to mean. Not “trust us, it feels real.” More like: constrain the generation, use formal scaffolding so the model is not just free-writing random emulator logic, then strategically test those behaviors against the actual cloud and patch the gaps when they show up. That is a much more serious answer than a lot of AI infrastructure demos give right now.

And honestly, that is where the episode got interesting for me.

Because if you are a platform engineer or DevOps person, you already know the pain here. Testing directly against real cloud is slow, expensive, and risky. Even when everything works, you are still paying in time, feedback delay, and blast radius. So the promise of something like Vera is not magic. It is faster iteration and safer validation. That is a much better frame for this than hype.

I also liked that Ang did not try to pretend the answer is perfect.

He says pretty directly that it is not one-to-one. The goal is not perfect imitation down to every line of output. The goal is to be close enough to support real classes of DevOps testing. I think that is the honest version of this whole category. Because if a sandbox can catch meaningful mistakes, break bad assumptions, and help validate changes before CI pushes something into actual cloud, that is already very valuable even if it is not a perfect clone of AWS.

The edge case he brought up was great too, because it shows how brutal infra tooling can be about details.

Something as dumb as camelCase versus snake_case in a response can be enough to break Terraform. That is the kind of thing people outside this space miss. Infrastructure tools are not impressed by “close enough.” They are extremely literal. So when people talk about cloud emulation, this is the real bar. Not whether it looks convincing in a demo. Whether it behaves precisely enough that existing tools do not choke on it.

Another part I liked was his answer on where this fits first.

Not everywhere. Not all at once. Plug it into CI/CD. Let it validate Terraform changes in a sandbox. Let it catch issues before push. That felt practical. At the same time, he was clear about limits too. EC2 was the main focus in the interview, it does not cover all AWS resources yet, and some of the more ambitious AI debugging and deployment-specific customization ideas are still on the roadmap. That honesty helps, because it keeps this grounded in “useful early tool” instead of “finished answer.”

The bigger thread running through the whole conversation is the one I keep coming back to.

AI for ops is probably not going to be won by whoever gives agents the most access. It is probably going to be won by whoever builds the best guardrails, the best evals, and the best places for those agents to learn safely. And that is what Vera feels like to me. Not the final form of AI in infrastructure, but a much smarter direction than pretending the path forward is just giving an LLM credentials and hoping for the best.

So if you are listening to this episode and want one takeaway, it is this:

Before AI earns the right to touch real infrastructure, it should have to survive a sandbox first.

That is the bar.

If you want, I can also tighten this into a slightly shorter, more spoken-word version for teleprompter delivery.

Show Notes

This is a guest conversation episode of Ship It Weekly, separate from the weekly news recaps.

In this Ship It: Conversations episode, I talk with Ang Chen from the University of Michigan about Project Vera, a cloud emulator built to help teams test infrastructure changes more safely before they touch real cloud.

We talk about why testing against real cloud APIs is slow, expensive, and risky, how Vera works under tools like Terraform and CloudFormation, what “high fidelity” actually means, and where a tool like this could fit in local dev and CI/CD.

The bigger theme is one I think matters a lot: if AI is going to play a real role in cloud operations, it probably needs a sandbox first, not direct access to production.

Note

This interview was recorded on February 13, 2026. Since then, Vera’s public project materials have expanded the framing a bit further around multi-cloud support and safe environments for agent learning, so keep that in mind while listening.

Highlights

• Why real cloud testing still creates cost, delay, and risk

• How Vera emulates cloud behavior at the API layer

• Where this could help with Terraform, CloudFormation, and CI/CD workflows

• Why “useful enough to catch real mistakes” may matter more than perfect emulation

• The limits, tradeoffs, and fidelity questions that still need to be solved

• Why safe training grounds may matter before AI agents touch real infrastructure

Ang’s links

• LinkedIn: https://www.linkedin.com/in/ang-chen-8b877a17/

• University of Michigan profile: https://eecs.engin.umich.edu/people/chen-ang/

• Publications: https://web.eecs.umich.edu/~chenang/pubs.html

Project Vera

• Project site: https://project-vera.github.io/

• GitHub: https://github.com/project-vera/vera

• The quest for AI Agents as DevOps: https://project-vera.github.io/blogs/cloudagent/cloudagent/

• No More Manual Mocks: https://project-vera.github.io/blogs/cloudemu/cloudemu/

Stuff mentioned

• A Case for Learned Cloud Emulators: https://dl.acm.org/doi/10.1145/3718958.3754799

• Cloud Infrastructure Management in the Age of AI Agents: https://dl.acm.org/doi/abs/10.1145/3759441.3759443

• LocalStack: https://www.localstack.cloud/

Our links

More episodes + show notes + links: https://shipitweekly.fm

On Call Brief: https://oncallbrief.com