Truth in IT
    • Sign In
    • Register
        • Videos
        • Channels
        • Pages
        • Galleries
        • News
        • Events
        • All
Truth in IT Truth in IT
  • Data Management ▼
    • Converged Infrastructure
    • DevOps
    • Networking
    • Storage
    • Virtualization
  • Cybersecurity ▼
    • Application Security
    • Backup & Recovery
    • Data Security
    • Identity & Access Management (IAM)
    • Zero Trust
    • Compliance & GRC
    • Endpoint Security
  • Cloud ▼
    • Hybrid Cloud
    • Private Cloud
    • Public Cloud
  • Webinar Library
  • TiPs

Scaling Terraform AWS Cloud Control Documentation with LLM

HashiCorp
04/06/2026
0
0 (0%)
Share
  • Comments
  • Download
  • Transcript
Report Like Favorite
  • Share/Embed
  • Email
Link
Embed

Transcript


TL;DR

  • The AWS Cloud Control provider automates Terraform resource generation weekly but lacked practical examples—700 resources existed with only 200 documented after two years of manual effort.
  • Early LLM experiments using long-context prompting produced inconsistent results due to attention loss and hallucinations from limited AWSCC training data.
  • Anthropic's Claude computer use enabled an agentic workflow where the LLM validates its own Terraform code through tool access, dramatically improving accuracy.
  • A production system using Step Functions and Lambda generated 450 working examples in three days for $400, nearly doubling two years of manual contributions.
  • The approach opens possibilities for pre-release validation, automated testing, continuous example maintenance, and potentially autonomous documentation updates.

The AWS Cloud Control Provider Documentation Challenge

The AWS Cloud Control (AWSCC) provider for Terraform offers automated weekly updates to support new AWS resources through the Cloud Control API, abstracting service-specific interactions into a consistent interface. However, this automation created a documentation gap: while the provider could automatically generate resource schemas, practical examples required manual creation. By 2023, the AWSCC provider included 700 resources, but only 200 had working examples despite two years of contributor effort. The challenge intensified as new AWS services launched faster than examples could be written, creating a growing backlog of undocumented resources that limited adoption.

Evolving the LLM Approach: From Prompts to Agentic Workflows

Initial experiments in early 2024 used large language models with extended context windows to generate examples from resource schemas and AWS documentation. While promising, this approach suffered from attention loss across long contexts and hallucinations due to limited AWSCC-specific training data. The breakthrough came with Anthropic's Claude computer use capability, which enabled an agentic workflow where the LLM could access tools like Terraform CLI, validate its own output, and iteratively refine examples. This shift from passive generation to active validation fundamentally changed the quality and reliability of generated documentation.

Production Architecture and Orchestration

The production implementation uses AWS Step Functions to orchestrate containerized Lambda functions, each handling specific phases: creation, validation, review, cleanup, and summarization. The system provides Claude with a secure, isolated environment containing Terraform binaries and access to resource schemas from the CloudFormation registry. System prompts establish governance rules around security best practices, tag formatting conventions, and resource-specific considerations like EKS cluster creation times. User prompts guide the LLM through sequential steps—downloading schemas, running terraform init and validate, applying configurations, and setting completion markers that trigger state transitions in the workflow.

Results and Future Applications

Over a three-day holiday period, the automated system generated 450 working resource examples at a cost of approximately $400 in Amazon Bedrock inference charges. This output nearly doubled the 250 examples created manually over two years by multiple contributors. The approach significantly reduced hallucinations by enabling the LLM to validate its own work through tool access and iterative refinement. Beyond documentation generation, the team identified potential applications including pre-release resource validation, automated testing of schema changes, continuous validation of existing examples as schemas evolve, and potentially autonomous pull request creation with appropriate human oversight.

Chapters

0:00 - Introduction and Speakers
1:19 - Why Two AWS Providers Exist
3:53 - The Documentation Problem
6:50 - First LLM Hypothesis and Limitations
8:32 - Anthropic Claude Computer Use
9:33 - Proof of Concept Demo
14:03 - Production Implementation Demo
16:59 - Architecture Deep Dive
21:40 - System and User Prompts
23:42 - Results and Impact
25:19 - Future Use Cases
26:24 - Closing and Resources

Key Quotes

1:55 "We have two AWS providers. It's called the AWS Standard Provider. And what's the new one? We typically call it AWS Cloud Control or AWS CC."
6:17 "We have like 700 resources in the AWS CC. We made very good progress. We added 200 examples since then. But then at the same time, new AWS services comes, new resources is now being implemented in Cloud Control API."
7:54 "At the time, I think, and still today, like a lot of this LLM has a very long contact length, 200k, right? Which is great, but often like you see the LLM lost the attentions."
9:06 "First, you give cloud or the LLM access to a tool and the prompt. The tool could be like Terraform binary. The tool could be a Python environment, bash and et cetera. And then cloud will decide what tool to use depending on the prompt."
15:18 "If you see the Terraform resource that is generated, it is not just generated the resource that we asked for, it generated all the subsequent or related supplemental resources that makes it a perfect, complete example for the user to use."
24:05 "For that three days, we are able to generate about 450 resources. And it take us about like 400 bucks in terms like the bedrock inference cost."
24:20 "Myself, Manu and couple other contributors working in early 2023 until end of the 2024. We have about 250 resources that we created and takes about two years."
25:03 "Human is always in the loop. We don't let the LLM to make a pull request. We still want to be the one for control. We supervise, look at the results."

Categories:
  • » Cybersecurity » Application Security
  • » Cybersecurity » Cloud Security
  • » Data Protection
Channels:
News:
Events:
Tags:
  • DevSecOps
  • AI & Machine Learning
  • Cloud Security
  • Technical Deep Dive
  • How-To
  • Terraform Provider Development
  • AWS Cloud Control API
  • LLM Agentic Workflows
  • Infrastructure as Code Documentation
  • Anthropic Claude Computer Use
  • Generative AI for DevOps
Show more Show less

Browse videos

  • Related
  • Featured
  • By date
  • Most viewed
  • Top rated
  •  

              Video's comments: Scaling Terraform AWS Cloud Control Documentation with LLM

              Upcoming Webinar Calendar

              • 04/08/2026
                11:00 AM
                04/08/2026
                Managing Configuration at Scale Across Group Policy and Intune
                https://www.truthinit.com/index.php/channel/1865/managing-configuration-at-scale-across-group-policy-and-intune/
              • 04/15/2026
                01:00 PM
                04/15/2026
                Service Account Security in the Age of AI: From Legacy Accounts to Agentic Identities
                https://www.truthinit.com/index.php/channel/1866/service-account-security-in-the-age-of-ai-from-legacy-accounts-to-agentic-identities/
              • 04/30/2026
                10:00 AM
                04/30/2026
                Insights from the 2026 Keepit Annual Data Report on SaaS Data Protection
                https://www.truthinit.com/index.php/channel/1868/insights-from-the-2026-keepit-annual-data-report-on-saas-data-protection/
              • 04/30/2026
                01:00 PM
                04/30/2026
                The New Economics of VMware Exit
                https://www.truthinit.com/index.php/channel/1880/the-new-economics-of-vmware-exit/

              Upcoming Events

              • Apr
                08

                Managing Configuration at Scale Across Group Policy and Intune

                04/08/202611:00 AM ET
                • Apr
                  15

                  Service Account Security in the Age of AI: From Legacy Accounts to Agentic Identities

                  04/15/202601:00 PM ET
                  • Apr
                    30

                    Insights from the 2026 Keepit Annual Data Report on SaaS Data Protection

                    04/30/202610:00 AM ET
                    • Apr
                      30

                      The New Economics of VMware Exit

                      04/30/202601:00 PM ET
                      More events
                      Truth in IT
                      • Sponsor
                      • About Us
                      • Terms of Service
                      • Privacy Policy
                      • Contact Us
                      • Preference Management
                      Desktop version
                      Standard version