Add blog post: AI coding agent orchestration frameworks compared#18493
Add blog post: AI coding agent orchestration frameworks compared#18493
Conversation
…compared New blog post comparing three community frameworks for AI coding agents. Covers Superpowers (TDD enforcement), GSD (context rot prevention), and GSTACK (role-based governance) with command reference tables, install instructions, and Pulumi workflow integration guidance.
Docs ReviewScope: All changed files in PR #18493 (blog post, SVG logos, feature/meta images) Issues
Minor Observations
Publishing Readiness Checklist
Mention @claude if you'd like additional reviews or automated fixes. |
Social Media Reviewcontent/blog/claude-code-orchestration-frameworks/index.mdTitle: Superpowers, GSD, and GSTACK: Picking the Right Framework for Your Coding Agent XDecision: FAIL Reasons:
Guidance:
Missing — not a failure, but noted for awareness. BlueskyMissing — not a failure, but noted for awareness. |
|
Your site preview for commit 83be317 is ready! 🎉 http://www-testing-pulumi-docs-origin-pr-18493-83be317d.s3-website.us-west-2.amazonaws.com |
Add linkedin and bluesky social keys to frontmatter. Add github-card shortcodes for the three framework repos in the getting started section.
Fix rule-of-three parallel structure, add personal opinion, vary framework descriptions, fix grammar in Twitter copy.
Social Media ReviewX — FAILReasons:
Guidance:
LinkedIn — FAILReasons:
Guidance:
Bluesky — PASSGuidance:
|
Address social-review feedback: replace "I tried" with neutral voice for corporate account, split Twitter into 2 paragraphs, add paragraph break to Bluesky for scannability.
|
@claude could you try rewriting the social posts so they pass and so they aren't summarizes of the article but instead a hook or setup for getting someone to read the article from the pulumi brand |
|
Claude finished @adamgordonbell's task in 2m 11s —— View job Rewriting Social Media Posts
Done. Here's what changed and why: The problem with the old copy: It summarized the article — listing which framework does what and telling readers that "GSD matters most for infrastructure." Once you've read that, you don't need to click. The new approach: Each post opens with the pain point readers already recognize, names the problem without naming the solution, and leaves the answer in the article. X/Twitter (passing — 2 paragraphs, no first person)
LinkedIn (3 paragraphs, no first person, professional tone)
Bluesky (2 paragraphs, no first person, hook)
· Branch |
Use Pulumi brand voice (no first-person), and open with the reader's pain point rather than the article's conclusions. Co-authored-by: Adam Gordon Bell <adamgordonbell@users.noreply.github.com>
|
@claude Rewrite the social posts to sound more natural. Remove AI-style phrasing, avoid em dashes, and use complete, uncluttered sentences. |
|
Claude finished @adamgordonbell's task in 1m 32s —— View job Rewriting Social Posts
Done. Here is what changed and why: Problems fixed:
X/Twitter
Bluesky
· Branch |
Co-authored-by: Adam Gordon Bell <adamgordonbell@users.noreply.github.com>
CamSoper
left a comment
There was a problem hiding this comment.
Approving — the post is in great shape, Engin. Star counts, install commands, and all 14 internal Pulumi links check out. Heading case, sentence case, and no em-dash issues.
One thing worth a look before you merge:
chardet attribution (Superpowers section). The "41x performance improvement" claim and chardet v7.0.0 both check out — the release notes literally say "41x faster than chardet 6.0.0 with mypyc, 28x pure Python". But the v7.0.0 release on GitHub was cut by dan-blanchard, not Jesse Vincent (obra). "Jesse used Superpowers to ship chardet v7.0.0" reads like Jesse is the maintainer who released it. If Jesse contributed the perf work via Superpowers and dan-blanchard merged/released it, something like "Jesse used Superpowers on his chardet contributions, which landed in v7.0.0 — 41x faster than 6.0.0 with mypyc, 28x in pure Python" would be more accurate. Worth a quick sanity check.
A couple of things I didn't verify (both likely fine, flagging in case you want to double-check):
- Lex Christopherson as GSD's creator (the repo is org-owned at
gsd-build, not directly attested on the landing page) - Garry Tan as YC CEO (widely known, didn't dig in)
Everything else is clean. The "know your failure mode" framing at the end is the right lens.
Dan Blanchard (chardet maintainer since 2012) used Superpowers to rewrite chardet v7.0.0. Jesse Vincent created Superpowers but did not write or release chardet. Fix attribution in both blog body and LinkedIn social copy.
|
Thanks Cam, good catch on the chardet attribution. You were right: dan-blanchard (chardet maintainer since 2012, 879 commits) used Claude Code with the obra/superpowers plugin to do the v7.0.0 rewrite. Jesse created Superpowers but did not write or release chardet. The Medium source material was ambiguous on this and I compounded it. Fixed in 3bfa13f: changed "Jesse used Superpowers to ship chardet v7.0.0" to "The chardet maintainer used Superpowers to rewrite chardet v7.0.0 from scratch." Same fix applied to the LinkedIn social copy (which the social-review bot had already rewritten, removing the reference). On the other two flags:
|
|
@claude update social posts in front matter: Suggested copy X (226/255 chars): We gave three AI coding frameworks the same Pulumi project. One produced a 41x speedup. One caught scope drift the others missed entirely. One we probably won't use again. Here's how they compared on real infrastructure work. We gave three open-source AI coding frameworks the same real infrastructure project and ran them for a few weeks. One produced a 41x speedup on a library release along the way. One caught a category of bug the other two missed entirely. One we probably won't reach for again. They have 270K combined GitHub stars and completely different theories about what goes wrong when agents work on longer projects. Turns out they're not all solving the same problem. Here's what we found and when to use which. We ran three AI coding frameworks on the same Pulumi infrastructure project for a few weeks. One caught scope drift the other two missed. One we probably won't use again. Here's how they compared. |
|
Claude finished @adamgordonbell's task in 40s —— View job Updating social media copy
Done. Replaced all three social posts in |
|
@dirien post looks great. Social post review process clearly struggled on this, but updates coming for that. This looks publishable to me. |
Co-authored-by: Adam Gordon Bell <adamgordonbell@users.noreply.github.com>
Docs ReviewOverall this is a well-structured, engaging comparison post with strong Pulumi tie-ins and good table-based comparisons. A few issues to address: Issues1. Style guide: prohibited words (lines 57, 153, 184) The style guide says to avoid words like "easy" or "simple." The post uses:
Consider replacing "straightforward" on lines 153 and 184 as well (e.g., "where unit tests are well-defined" and "the install commands are short"). 2. First mention of tools not hyperlinked (line 37) Per blog review criteria, the first mention of every tool, technology, or product must be hyperlinked. Cursor, Codex, Windsurf, and Gemini CLI are all mentioned in the first paragraph without links. 3. Social copy sets expectations the post doesn't deliver (frontmatter lines 19-33) All three social blocks say "One we probably won't use again," which creates a strong negative verdict. The blog itself is balanced and never identifies a framework it would not use again. Readers who click through expecting a clear winner/loser will feel misled. Either soften the social copy or add a clearer verdict in the post. 4. "S3 bucket" link destination is generic (line 47) The anchor text "S3 bucket" links to 5. LLM acronym not expanded at first use (line 47) "Every LLM has a context window" — while widely known in AI circles, the style guide requires non-Pulumi acronyms to be spelled out on first use. Consider "Every large language model (LLM) has a context window" at the first occurrence. 6. Star counts will go stale quickly (lines 57, 84, 112, 148) The post cites specific GitHub star counts (149K, 51K, 71K) for each framework. These change rapidly. Consider adding "at time of writing" or using approximate language, or removing the exact counts from the comparison table since the Minor notes
Publishing readiness checklist
Mention @claude if you'd like additional reviews or automated fixes. |
Social Media Reviewcontent/blog/claude-code-orchestration-frameworks/index.mdX — PASSLinkedIn — PASSBluesky — PASSTo apply these suggestions, comment: To re-run the social media review after updates, comment: |
|
📣 Social media post results for this blog post:
|
|
📣 Social media post results for this blog post:
|
|
📣 Social media post results for this blog post:
|
Summary
Test plan
/blog/claude-code-orchestration-frameworks//docs-reviewfor style compliance