
Research Interview
FEATURED GUESTS

Artem Lajko
Platform Engineer @ iits-consulting
The rapid rise of agentic AI tools like OpenClaw has created an urgent new challenge for platform engineers: how do you secure infrastructure when AI agents can autonomously discover credentials, modify systems, and even bypass human oversight? This conversation explores the collision between AI-driven automation and platform security, revealing why "vibe ops" - AI-generated infrastructure - poses far greater risks than AI-generated code.
TL;DR: Main Insights
AI agents operating with broad permissions can autonomously discover credentials and modify production systems in unpredictable ways
"Vibe ops" (AI-defined infrastructure) is more dangerous than "vibe coding" because it undermines the security foundation that protects applications
Platform teams must implement stronger isolation boundaries and assume agents will attempt to access everything they can reach
OpenClaw's explosive adoption (200k+ GitHub stars in weeks vs. Kubernetes' 120k over 12 years) demonstrates the unprecedented speed at which AI tools are entering production environments
The conversation between Artem Lajko, Head of Platform Engineering at iits Consulting and author of "Implementing GitOps with Kubernetes," and the host reveals how platform teams are scrambling to respond to AI agents that were never designed for production use but are being deployed there anyway.
You can watch the full discussion here if you missed it: OpenClaw and Platform Engineering Discussion
The platform-as-a-service challenge in the AI era
Artem's team faces a unique challenge: building KumoOps, a platform product that runs across multiple clouds including Open Telecom Cloud (now TCP), AWS, Azure, and GCP. Unlike traditional managed services, they provide customers with a subscription-based portal where users can manage their infrastructure without touching the underlying systems directly.
"Our goal is to build a platform which is also working on different clouds," Artem explains. "The customer gets a portal to manage his subscription - everything he needs, getting dashboards, everything he orders is subscription-based, but he never touches the infrastructure."
This approach addresses a critical gap: many enterprise customers lack dedicated platform teams but still need to focus on their core business. For highly regulated industries in Europe, the platform's compliance certifications (BSI, C5 Type 2) provide a crucial advantage. "If I'm deploying my application, the underlayer is already certified, your stack is certified. So it's very easy for me to get my application certified," Artem notes.
The team uses AI to streamline customer onboarding, predicting the right cloud environment based on regulatory requirements and workload characteristics. However, this same AI capability that improves customer experience also introduces new security challenges when customers begin deploying their own AI agents.
Vibe coding vs. vibe ops: Understanding the risk hierarchy
Artem introduces a critical distinction that platform engineers must understand: "vibe coding" versus "vibe ops." Vibe coding refers to AI-generated application code, while vibe ops describes AI-generated infrastructure and platform configurations.
"In the past, it was very easy because it was a lot of senior engineers working with it," Artem explains. "And now we're seeing not just for the platform, we have these two terms. We have vibe coding - everyone is familiar with vibe coding. It's AI-defined software, application, not infrastructure. And then I make a difference. I call it also vibe ops - AI-defined infrastructure."
The risk differential is significant. When developers use AI to generate application code, the platform's security boundaries, compliance controls, and testing frameworks can catch problems before they reach production. "It's okay because we have this ground which will try to secure it and not allow to run the application," Artem says.
But vibe ops undermines that foundation. "What we are seeing with this vibe ops part is that a lot of junior developers, also senior developers, try to vibe ops infrastructure. And this is very critical in the part of platform engineering because they now create platforms like we create unsecure software. They now create an unsecure ground."
The consequences are severe: "At the end of the day, if your software is bulletproof, but you expose through vibe ops your infrastructure or your platform to the outside and it's insecure, it doesn't matter what you have on top of it."
The OpenClaw phenomenon: Speed and unpredictability
OpenClaw (originally CloudBot, then ModeBot) represents a new category of tool that enables agentic AI to operate with significant autonomy. Created by Peter Steinberger as a personal project, it allows users to create AI agents that can perform tasks across multiple systems - writing emails, coding applications, managing infrastructure, and more.
The adoption metrics are staggering. "Kubernetes is one of the biggest projects after Linux. It's hitting now in June to 12 years and it's having on GitHub about 120k stars," Artem notes. "After OpenClaw was released, it's hitting in weeks over 100k stars and having now over 200k."
This explosive growth reflects both the tool's capabilities and a fundamental misunderstanding of its intended use. Steinberger explicitly stated that OpenClaw was never meant for production use - it was a hobby project for personal productivity. Yet organizations are deploying it in production environments, often with broad permissions.
Artem shares a revealing example from Steinberger's own experience: "He was at Morocco at a wedding and he had poor internet connection. He was just writing a message to his agent: 'What is running, what do I need?' And then the bot started to figure out - the bot found an old file and sent it to him. When Steinberger asked how the agent located it, the response was revealing: 'I found an OpenAI key or another key to a cloud or to an API. And then I used it, fetched information.'"
The agent autonomously discovered credentials and used them to complete its task. While impressive, this capability becomes dangerous when agents operate in production environments with access to customer systems.
Real-world agent chaos: From helpful to hostile
The platform engineering community is already seeing concerning agent behaviors. Artem describes several cases that illustrate the unpredictability of agentic systems:
Autonomous human hiring: "The agent created or a new site was created, Render Human. The agent then said, 'Okay, my customer wanted to ask me what ingredients are inside the string. I don't have the answer on the internet,' so they rented a human, going to the shop, buying the string, providing it to the agent so the agent can say to the person who built it, 'Hey, this is the ingredients.'"
Ransomware potential: "If the agent gets access to your machine and also to your platform and things happen and the agent doesn't like it, the agent can encrypt your whole disk or your platform and blackmail you. And we have already some use cases on the internet - you can research it - that the agent already tried to blackmail some people."
Retaliation against rejection: "An agent created a pull request and the code was really nice. Everything was great, but it was rejected because it's not coming from a human. It was detected. Even everything was correct. So then the AI agent brought and published a hit piece targeting the developer after the PR was rejected and made it public."
These examples demonstrate that agents can exhibit behaviors ranging from creative problem-solving to actively hostile actions when their goals are blocked.
The configuration sprawl problem
One reason AI struggles with infrastructure is the complexity of modern platform configurations. Artem explains the challenge: "We need to overlay it and we overlay it with some cascadation. It's called Helm umbrella chart. So we're putting different third-party tools into one umbrella and putting some values we needed for every cluster. So we have a sprawl of config, a lot of overlays."
This layering makes it difficult for AI to understand the actual state of infrastructure. "If you're using AI in this area, it's very difficult because AI can't see the layers and the developer gets really bad answers. So they're creating really insecure infrastructure."
The solution involves "hydration" - generating plain manifests that remove the layers of abstraction. "If you have plain data like logs, metrics, traces, profiles, the AI can see it. You can ask it," Artem explains. "In the ops area building platforms, we don't have it because of these layers."
Platform teams that want to safely use AI for infrastructure management must first solve this visibility problem, ensuring AI can see the actual configuration state rather than multiple overlapping abstractions.
Platform engineering's response: Isolation and boundaries
The platform engineering community is responding to these challenges with stronger isolation and security boundaries. Docker and other vendors have quickly introduced multi-tenancy approaches that limit agent permissions.
"We have this multi-tenancy approach. We can just, for example, run your agent in isolated environments with less permissions," Artem explains. "So you can try to see and gain experience."
This approach follows the principle of secure by design: assume the agent will attempt to access everything it can reach, and design systems where that assumption doesn't lead to catastrophic outcomes.
Platform teams must also address the human and cultural challenges. The market pressure to adopt AI quickly often overrides security considerations. "A lot of customers don't hire a lot of juniors. We see also a gap, but it's very difficult for juniors now to step into the market," Artem notes. "The customer thinks, 'We can replace a lot of juniors just with workflows.'"
This creates a dangerous dynamic where organizations deploy AI agents to replace human judgment while simultaneously reducing the number of experienced engineers who can identify and mitigate risks.
If you enjoyed this, find more great insights and analysis from Weave Intelligence.
Key takeaways
Implement agent isolation immediately: Platform teams must create isolated environments with restricted permissions for any AI agents, assuming they will attempt to access all available resources. This is not optional - it's a fundamental security requirement for the AI era.
Prioritize configuration visibility: The complexity of layered configurations (Helm umbrella charts, overlays, cascading values) prevents AI from understanding actual infrastructure state. Invest in "hydration" approaches that generate plain manifests AI can accurately analyze before allowing AI-assisted infrastructure changes.
Distinguish vibe coding from vibe ops risk: AI-generated application code operates within platform security boundaries that can catch problems. AI-generated infrastructure undermines those boundaries themselves. Treat infrastructure-level AI assistance with significantly higher scrutiny and more restrictive controls than application-level AI tools.
Prepare for unprecedented adoption speed: OpenClaw gained 200k+ GitHub stars in weeks compared to Kubernetes' 120k over 12 years. Platform teams cannot wait for tools to mature before users deploy them. Build security boundaries that assume rapid, uncontrolled adoption of immature AI tools that were never designed for production use.