Your Style Guide Has 200 Rules. Your Information Developers Remember 12. How to Fix that?
This blog post describes a self-review skill built on GitHub Copilot's extensibility framework. The approach and architecture described here are generalizable to any domain where editorial or compliance standards can be expressed as structured rules and applied by the content creator before formal review.
Technical documentation is hard to keep consistent. Style guides grow to dozens of pages. Terminology tables expand with every release. New information developers join, experienced information developers forget edge cases, and everyone submits drafts with issues they could have caught themselves—if they had the time and mental bandwidth to cross-check every rule.
Our team maintains documentation for a mature enterprise platform with strict editorial standards—over 14 broader categories of review criteria, from terminology accuracy and grammatical correctness to structural completeness and typographic conventions. Information developers genuinely wanted to get things right the first time, but no one can hold hundreds of rules in their head while also thinking about clarity, accuracy, and structure. The result was predictable: editors spent most of their time catching mechanical errors instead of focusing on the high-value work—content accuracy, narrative flow, and reader experience.
We asked ourselves: What if information developers had a way to self-review against the full style guide before submitting a draft?
The core insight was simple—our editorial standards were already codified in a written style guide. The rules were explicit, the examples were concrete, and the expected corrections were well-defined. This is exactly the kind of structured knowledge that a large language model can apply consistently, at scale, and without fatigue. The missing piece wasn't the rules—it was putting them in the information developer's hands at the moment of writing.
Rather than building a standalone application or another review gate, we chose to build a GitHub Copilot custom skill—a modular extension that plugs directly into the information developer's IDE. This meant:
The skill is remarkably lightweight. The entire system consists of three core files:
This is the orchestration layer—a structured Markdown file that tells the AI who it is and how to behave. It defines:
The key design principle here is deterministic orchestration with flexible execution. The steps are numbered and mandatory but the AI has latitude in how it applies each rule to the specific document at hand. From the information developer's perspective, it feels like having a tireless colleague who has memorized the entire style guide and is always available for a quick read-through.
This is the knowledge base—the encoded style guide. It contains:
The criteria file is loaded by reference during the review—the AI reads it at runtime, which means updates to the style guide are immediately reflected in reviews without redeploying anything.
This is the output schema—a structured Markdown template means that every review produces consistent, machine-readable output:
The template enforces a contract: every review looks the same, regardless of which document was reviewed or when.
For a self-review tool to earn information developers' trust, it has to be precise—not vague. The most important decision we made was to express every editorial rule as a concrete, testable criterion with examples of both correct and incorrect usage. Instead of "use clear language," we wrote specific tables:
Flag this | Replace with |
“in order to” | “to” |
“has the ability to” | “can” |
“prior to” | “before” |
This precision reduces hallucination and subjective variation. The AI isn't deciding what "clear" means—it's pattern-matching against an explicit inventory of known issues. For information developers, this is crucial: the feedback feels objective and actionable, not like an opinionated critique.
Every issue is classified into one of three severity levels:
This three-tier system gives information developers a clear prioritization framework during self-review. A document with twenty Minor issues and zero Critical issues is in much better shape than one with two Critical issues and the severity labels make that immediately visible. Information developers can focus on the Critical and Major items first and address Minor polish if time allows — just as a human editor would advise.
The skill doesn't just find issues—it computes measurable document health indicators:
These metrics give information developers a numeric snapshot of their draft's health before they even read the detailed findings. A quick glance at the summary answers the question every information developer asks before submitting: "Is this ready or does it need another pass?" The metrics also make it possible for information developers to track their own improvement over time.
Not all documents are the same. A courseware lesson has learning objectives and check-your-understanding questions; a reference topic has syntax blocks and parameter tables; release notes have neither.
The skill auto-detects the document type from structural signals (headings like "Learning Objectives" or "Syntax" or "Release Notes") and activates or deactivates entire categories of checks accordingly. Courseware gets Bloom's Taxonomy verb validation and objective-assessment alignment checks. Reference docs get syntax-example pairing checks. Release notes get only the universal grammar, style and terminology checks.
This conditional architecture prevents false positives—the skill never flags a reference topic for missing learning objectives.
Every review produces the same output structure:
The consistency of the output format means that information developers learn the layout once and can scan it quickly on every subsequent self-review. When they do submit to an editor, the editor receives a cleaner draft—and if the information developer attaches the self-review report, the editor can see exactly what was already checked.
One of the more novel aspects of the design is how the skill blends LLM reasoning with deterministic computation. The AI doesn't estimate passive voice percentage—it runs a Python script that counts passive constructions using regex patterns and NLP part-of-speech tagging.
This hybrid approach plays to each technology's strengths:
The result is a review that is both computationally rigorous and editorially intelligent.
Since deploying the skill as a self-review tool:
You don't need our specific domain or our specific platform. The pattern generalizes to any team that has editorial standards and produces written content. The goal is simple: give every information developer the ability to review their own work against the team's full set of standards, instantly and on demand. Here's the playbook:
Step 1: Audit Your Style Guide
Go through your existing style guide and classify every rule into one of these types:
Step 2: Build the Criteria Reference
Create a single document that contains all your rules in explicit, example-rich format. Structure it with clear section numbers so the AI can cite specific rules in its justifications. Include both positive examples (correct) and negative examples (incorrect) for every rule.
Step 3: Define the Output Contract
Design your annotation template before you build the skill. Decide:
Step 4: Write the Skill Definition
Define the persona, the procedure and the input/output handling. Be specific about the order of operations. Include error handling for edge cases (PDF extraction failures, ambiguous document types, etc.).
Step 5: Add Computational Metrics
Identify which aspects of quality can be measured numerically. Sentence length, passive voice density and acronym coverage are good starting points. Write small scripts that compute these metrics and integrate them into the review flow.
Step 6: Iterate With Real Documents
Run the skill against real documents from your team. Have information developers use it for self-review and collect their feedback. Tune the criteria based on false positives and missed issues. Add exceptions for legitimate edge cases. Refine the severity classifications based on what information developers actually find actionable. The skill improves fastest when the people using it daily are the ones shaping the rules.
Start with the rules, not the AI. The quality of the skill is directly proportional to the quality of the encoded criteria. Spend most of your time on the style guide, not the prompt engineering.
Be explicit about exceptions. Every terminology rule has edge cases. Document them upfront. Without explicit exceptions, the AI will flag legitimate usage and information developers will stop trusting the tool. Trust is everything in a self-review system—the moment information developers dismiss the output as noisy, they stop using it.
Severity matters more than count. A review that finds 50 Minor issues is less actionable than one that finds 3 Critical issues. The severity framework is what transforms a wall of feedback into a prioritized to-do list that information developers can actually work through.
Self-review is not a replacement for editorial review. The skill catches mechanical and structural issues — the things that are tedious for humans but straightforward for a machine. It does not replace the editor's judgment on content accuracy, audience appropriateness or narrative quality. What it does is free up the editor's time and attention for exactly that high-value work.
The skill teaches the style guide. Every self-review is a micro-learning moment. The justification column explains why each change is recommended, citing specific rules. Over time, information developers internalize the standards and make fewer errors in the first draft. The skill doesn't just catch mistakes—it prevents them.
Structured output enables automation. Because the output follows a consistent schema, downstream processes can consume it—tracking metrics over time, generating quality dashboards or gating submissions on minimum quality thresholds.
The skill is the style guide. Once the criteria are encoded in a machine-readable format, the "style guide" is no longer a document that people read once during onboarding—it's a system that actively coaches information developers every time they self-review. Updates to the criteria file immediately change the behavior of every future review.
This is just the beginning. The same pattern—encode domain expertise as structured criteria, let the LLM apply it with judgment and augment with deterministic computation—can be applied as a self-review tool beyond documentation:
The common thread is shifting quality left—empowering the person doing the work to catch issues at the point of creation, not the point of review. The tools are here. The pattern is proven. The only prerequisite is a team that has codified its standards—and the willingness to put those standards directly in the hands of the people who write.
Senior Information Developer, OpenEdge Content and Courseware Team
Deepti Gupta has over 16 years of experience spanning software development, content development and technical writing. Her career began as an Oracle Apps Developer, where she built a strong foundation in ERP databases, business reports and supporting financial operations through Oracle Apps Finance modules.
This deep technical background now informs her approach to documentation, ensuring clarity in every piece of technical and courseware content she creates. Eager to implement AI-enhanced workflows, Deepti seeks to learn about the broader applicability of AI across technical documentation and instructional design, including adaptive learning paths and microlearning modules.
Driven by curiosity and a commitment to continuous improvement, Deepti advocates for hybrid workflows that combine AI efficiency with human creativity. Her goal is to empower teams to produce high-impact content that resonates with users and supports business goals. In her free time, she enjoys reading literary fiction, learning to play the ukulele and watching mind-bending sci-fi movies.
Subscribe to get all the news, info and tutorials you need to build better business apps and sites