Back to blog
May 23, 2025
13 min read

Presales: GitOps Transformation Technical Validation - 95% Maintenance Reduction

How I achieved the technical win for a GitOps solution using Terraform, ArgoCD, and HashiCorp Vault, demonstrating 95% infrastructure maintenance reduction through technical validation

Presales: GitOps Transformation Technical Validation - 95% Maintenance Reduction

The Presales Opportunity

This opportunity came through our Account Executive during a quarterly business review with an existing client. While we had a footprint with their payment processing platform, their DevOps team was struggling with infrastructure management bottlenecks that were impacting their ability to scale.

The situation was fairly straightforward - they recognized the need to modernize their infrastructure management but were unsure about the best path forward. They weren’t evaluating multiple vendors in a formal RFP process, which made things both easier and harder. Easier because there wasn’t competitive pressure, but harder because there wasn’t urgency driving a decision timeline.

The DevOps Director, Sarah, had been researching different Infrastructure as Code options including AWS CDK, Pulumi, and Terraform. There was also the option of doing nothing and continuing with their current manual processes.

Sarah was the technical buyer - she could definitely say no to any proposed solution, but the final budget decision would rest with the CTO. She was naturally skeptical and preferred a “show me, don’t tell me” approach.

Technical Discovery: Understanding the Real Problems

My first step was conducting thorough technical discovery to understand their current pain points and requirements. I’ve learned that you can’t design an effective solution without really understanding the problem, and often the stated problem isn’t the real problem.

Their setup was typical of a growing company - they had 35 microservices running across development, staging, and production environments, deployed across both AWS and Google Cloud Platform. Everything was managed manually through SSH sessions, configuration file edits, and manual restart sequences. While this had worked when they were smaller, it was becoming a real bottleneck as they scaled.

The multi-cloud setup added complexity to their infrastructure management. They had chosen AWS initially but later expanded into GCP for specific services and data locality requirements. Managing infrastructure across two cloud providers manually was becoming increasingly cumbersome, especially when they needed to ensure consistent security policies and networking configurations across both platforms.

The most time-consuming issue was their infrastructure maintenance overhead. Sarah’s team was spending about 32 hours per week on manual infrastructure tasks - that’s almost a full-time person just keeping the lights on. This included things like manually updating certificates, applying configuration changes across environments, and troubleshooting deployment issues that often behaved differently between AWS and GCP.

Configuration drift was another major pain point. Their production, staging, and development environments had slowly diverged over time as quick fixes and manual changes accumulated. This meant that deployments often worked in staging but failed in production, leading to late-night debugging sessions and delayed releases. The multi-cloud aspect made this even worse, as they had to maintain consistency across different cloud providers with different tooling and APIs.

From a security perspective, they had inconsistent certificate management across their 35 services. Some certificates were managed manually and others were handled by different automated processes. Certificate expiration had already caused a few production incidents, and they knew this was going to get worse as they scaled.

The compliance aspect was becoming increasingly important as well. They needed audit trails for PCI DSS and SOC 2 compliance, but their manual processes made it difficult to track who changed what and when. Sarah mentioned that their last compliance audit had taken weeks of manual documentation gathering.

Solution Design: Building on Existing Investment

Based on what I learned in discovery, I designed a GitOps solution using Terraform for infrastructure management, ArgoCD for application deployments, and HashiCorp Vault for centralized secrets and certificate management.

Terraform was the natural choice for several reasons. Their team was already most familiar with Terraform among the Infrastructure as Code options they were considering. Sarah’s team had been experimenting with Terraform modules for some of their AWS infrastructure, so there was existing knowledge to build on. Terraform is the most mature and widely adopted IaC tool in the market, with extensive provider support. Most importantly, Terraform’s multi-cloud capabilities would allow them to manage both their AWS and GCP infrastructure from a single toolset with consistent workflows.

For their certificate management challenges, Vault PKI would act as a centralized certificate authority. This would handle automatic certificate generation and rotation for all 35 services, eliminating the manual overhead and reducing the risk of certificate-related outages.

The GitOps workflow would address their configuration drift problem by making Git the single source of truth for both infrastructure and application state. Any changes would go through code review and be automatically applied, ensuring consistency across all environments.

Technical Validation: Demo and Proof of Value

Since Sarah was skeptical about “yet another tool,” I knew a standard sales demo wouldn’t be sufficient. She needed to see the solution working with their actual infrastructure and requirements.

I started with a customized demo that replicated their environment - 35 mock services with similar communication patterns. The demo showed the entire GitOps workflow from a developer making a change in Git to that change being automatically deployed across environments.

The Vault PKI integration was the centerpiece of the demo. I showed how certificates would be automatically generated and rotated for service-to-service communication, with zero manual intervention required. This directly addressed one of their biggest operational pain points.

The demo went well, but Sarah’s main concern was about the complexity of integrating with their existing infrastructure. She worried about disrupting production services and the learning curve for her team. This is where I suggested a targeted Proof of Value using a subset of their actual services.

The Skeptical Champion Challenge

During our second technical meeting to discuss the PoV scope, I encountered a situation I’ve seen many times before. One of Sarah’s senior engineers, let’s call him Marcus, started raising concerns about the proposed solution. About 15 minutes into the meeting, he began questioning whether GitOps was really necessary, suggesting they could achieve similar results with some bash scripts and better documentation.

I’ve learned from experience that these moments can make or break a deal, but they’re also opportunities if you handle them correctly. The key is having a clear meeting agenda that everyone agrees to at the start, and then not letting anyone hijack the discussion.

When Marcus started going down this rabbit hole, I politely interrupted and said, “Marcus, these are really good questions and I want to make sure we give them the attention they deserve. But right now we’re focused on defining the PoV scope according to our agenda. Could we schedule some time after this call to dive deeper into your concerns?”

This is a technique I’ve refined over years of presales meetings. When you have a skeptical techie, you need to determine quickly whether they’re a potential champion or a detractor. If they’re genuinely interested in understanding the solution, they’ll take you up on that follow-up conversation. If they’re just being contrarian, they’ll usually decline and move on.

Marcus agreed to the follow-up call, which told me he was likely a potential champion rather than a detractor. Two days later, we had a 45-minute technical deep dive where I could address his specific concerns.

His main objection was complexity - he felt like they were adding too many moving parts when simpler solutions might work. I walked him through the architecture step by step, showing how each component solved a specific problem they currently had. When he suggested bash scripts could handle deployments, I showed him how ArgoCD’s declarative approach eliminates the need to write and maintain deployment scripts while providing better visibility and rollback capabilities.

Another concern he raised was about Terraform versus AWS CDK and Pulumi. I explained that while CDK would work well for their AWS resources, it wouldn’t help with their GCP infrastructure, forcing them to maintain different toolsets and workflows for each cloud provider. Terraform’s multi-cloud approach would let them use consistent patterns and modules across both AWS and GCP, which was especially valuable given their team’s existing Terraform experience.

The turning point came when we discussed certificate management. Marcus admitted they’d had three production incidents in the past year related to expired certificates, and he was personally spending about 4 hours a month just tracking certificate expiration dates across both cloud environments. When I showed him how Vault PKI would completely eliminate this operational overhead and work consistently across AWS and GCP, he got visibly excited.

By the end of the call, Marcus had shifted from skeptic to advocate. He even volunteered to be the technical point of contact for the PoV implementation. Technical people who ask hard questions often become your strongest champions once they understand the solution.

Proof of Value Implementation

We agreed on a focused PoV that would validate the most critical aspects of the solution. The scope included 5 of their 35 microservices in their staging environment, with the goal of demonstrating complete deployment automation and automated mTLS certificate management.

The implementation took about three weeks. The most challenging part was configuring Vault to work as the certificate authority for their existing service mesh setup. This required careful coordination to avoid disrupting any running services.

Marcus was instrumental during this phase. Having a technical advocate inside the customer organization made everything smoother - he understood their infrastructure better than I did and could make configuration decisions quickly. He also helped communicate progress to Sarah and address any concerns from other team members.

The results were compelling. We demonstrated a complete GitOps workflow where code changes automatically triggered infrastructure updates and application deployments. Certificate rotation was happening automatically every 24 hours with zero service disruption. Most importantly, what used to take 45 minutes of manual work per service deployment now took 3 minutes of automated process.

Achieving the Technical Win

The PoV results spoke for themselves. We had quantifiable improvements in deployment speed, operational overhead reduction, and security posture. But the real technical win came from having Marcus as an internal champion who could articulate the benefits to his peers.

During the final presentation, Marcus walked through the technical implementation with Sarah’s team, explaining how each component worked and addressing questions from his colleagues. This was much more credible than if I had presented the same information as an external vendor.

Sarah’s decision was straightforward: “This solves our operational overhead problem, improves our security posture, and builds on tools we already know. Let’s move forward.” The projected 95% reduction in infrastructure maintenance time - from 32 hours per week to about 1.5 hours - provided clear business justification for the investment.

The key to this technical win was understanding that it wasn’t really about the technology. It was about proving that we could deliver measurable operational improvements without disrupting their existing systems or overwhelming their team with complexity. By turning a skeptic into a champion and demonstrating real results with their actual infrastructure, we eliminated the technical risk that often kills these deals.

From a presales perspective, this engagement reinforced the importance of thorough technical discovery, managing meeting dynamics to turn potential objections into advocacy opportunities, and proving solutions work in the customer’s specific environment rather than just in demos.

  • Familiar HashiCorp tooling reduced learning curve

4. Compliance and Audit:

  • Complete audit trail through Git commits and Terraform state
  • Policy as code for automated compliance validation
  • Immutable infrastructure eliminating configuration drift

The Technical Win

After the PoV presentation, Sarah’s feedback was decisive: “This is exactly what we need. The Vault integration alone saves us hundreds of hours annually, and the fact that it works with our existing Consul setup means minimal disruption to our current operations.”

Key factors that secured the technical win:

  1. Proven Integration: The PoV demonstrated seamless integration with their existing HashiCorp stack
  2. Quantified Benefits: Clear metrics showing 95% maintenance reduction
  3. Risk Mitigation: Leveraged existing tools rather than requiring wholesale changes
  4. Security Enhancement: Vault PKI solved their service-to-service communication challenges
  5. Audit Capability: Git-based audit trail addressed compliance requirements

Handoff to Implementation

With the technical win secured, my role shifted to ensuring smooth handoff to the implementation team:

Implementation Readiness:

  • Detailed architecture documentation from the PoV
  • Terraform modules ready for production deployment
  • Team training plan for HashiCorp tooling
  • Migration strategy for remaining 30 services
  • Success metrics and monitoring plan

Commercial Handoff: The technical validation de-risked the commercial negotiation. With clear evidence of the solution’s effectiveness, our Account Executive could focus on contract terms and expansion opportunities rather than justifying the technical approach.

Professional Services Engagement: Based on the PoV complexity, we recommended a 6-week professional services engagement to:

  • Migrate remaining 30 services to GitOps workflow
  • Complete Vault PKI rollout across all environments
  • Train their team on advanced HashiCorp features
  • Establish monitoring and alerting for the new infrastructure

Presales Challenges and Key Lessons

The most challenging aspect was configuring Vault PKI integration, which took longer than expected. Sarah’s team also wanted to expand the PoV scope beyond what we had agreed, so I had to maintain focus while documenting expansion items for the implementation phase.

Key success factors were focusing on their biggest pain point (manual certificate management across 35 services), having concrete time measurements (45 minutes → 3 minutes deployment time), and implementing the solution in their actual staging environment rather than just showing demos.

The 4-week PoV investment secured a 6-figure deal plus professional services engagement. More importantly, the detailed architecture documentation became a reusable asset for similar opportunities, significantly reducing time-to-technical-validation on future deals.

Conclusion

This presales engagement demonstrates how thorough technical validation can secure the technical win. The 95% reduction in infrastructure maintenance wasn’t just a metric - it was the quantified business outcome that convinced the technical buyer.

Key takeaways: solve the biggest pain point first, measure everything with concrete time savings, implement solutions in the customer’s actual environment rather than isolated demos, and turn skeptics into champions through detailed technical discussions.

The technical validation won the specific deal and established a reusable playbook for similar opportunities, demonstrating the value of HashiCorp’s integrated platform approach for enterprise infrastructure modernization.