/mktg:expert-panelExpert Panel
Assemble un panel de 7-10 experts adaptés au type de contenu (copywriter senior, audience cible, CRO specialist, designer, fondateur sceptique...). Chaque expert note /10 et livre ses 3 fixes prioritaires avec citation. Le Humanizer (détecteur de slop IA) est toujours inclus à 1.5x de poids. Si le score est < 90, itère un round de rewrite puis re-score. Stoppe quand score >= 90 ou après 3 rounds.
Marketers, fondateurs, product leads qui veulent passer d'un draft acceptable à un texte excellent avant un lancement. Compte 5-10 min par round et jusqu'à 3 rounds. Pas fait pour du contenu de routine : à réserver aux textes où la qualité compte vraiment (landing, pitch, annonce, page critique).
Configuration
mktg:expert-panelReadWriteWebFetch<text, file path, or URL> [content type]Installation
$ mkdir -p ~/.claude/skills/mktg:expert-panel && \
curl -sSL https://www.cedricrittie.com/api/skills/mktg-expert-panel/raw \
-o ~/.claude/skills/mktg:expert-panel/SKILL.mdPuis redémarre Claude Code. Test avec `/mktg:expert-panel`.
Besoin d'installer Claude Code d'abord ? Voir la fiche Claude Code · Télécharger le .md brut
Exemples d'utilisation
/mktg:expert-panel article.mdPanel auto-assemblé pour article. Round 1 : scoring + fixes. Si < 90, rewrite intégré, puis round 2.
Panel assemblé (8 experts) - Copywriter senior - Audience cible (PM avec 5+ ans d'expérience) - Spécialiste CRO - Rédactrice en chef - Fondateur sceptique - Humanizer (poids 1.5x) - Journaliste - Designer Round 1 — Score : 74/100 Problèmes principaux 1. Copywriter senior (6/10) : l'attaque enterre l'insight 2. Humanizer (5/10) : 7 em-dashes, 2 "let's dive in" 3. Audience cible (7/10) : jargon au paragraphe 3 [...] Rewrite appliqué. Round 2 — Score : 91/100 ✓ Prêt à publier. Feedback complet dans ~/skills/mktg:expert-panel/last-run.md
/mktg:expert-panel https://landing.com landing-pagePanel spécialisé CRO pour landing. Différent du panel article : CRO Specialist prend plus de poids, nouvelles dimensions évaluées.
Panel assemblé — spécialistes landing page [...] Round 1 — Score : 68/100 Prêt à publier après round 3 — Score : 92/100
Le Skill en entier
Expert Panel — Iterative Multi-Expert Scoring
Takes any content and scores it through a panel of simulated domain experts. Loops until the score hits 90+ or 3 rounds are exhausted. The humanizer expert is always included at 1.5x weight.
Input
Accept content in any of these forms:
- Pasted directly as the argument
- A file path -> read the file
- A URL -> fetch the content
- Content type hint (optional): landing-page, email, social-post, article, pr, strategy, ad-copy
If no content type is specified, infer from the content itself.
Step 1: Intake
Identify:
- Content type: what are we scoring?
- Goal: what should this content achieve?
- Audience: who reads this?
- Context: any constraints (brand voice, platform limits, regulations)?
State these in 4 lines max before proceeding.
Step 2: Auto-Assemble the Panel
Select 7-10 experts based on content type and domain. Always include:
- Humanizer (mandatory, weight 1.5x) -- Detects AI writing patterns. Scores how human the text sounds. Uses the 24-pattern detection framework from writing:unslop.
- Brand Voice Match (mandatory) -- Does this sound like the brand/person, or like generic AI output?
Then add 5-8 domain experts from the pre-built panels below.
Landing Page
- Conversion Rate Optimizer (CTA, friction, clarity)
- Headline Specialist (hook, specificity, scroll-stopping)
- Social Proof Analyst (trust signals, testimonials, credibility)
- UX/Mobile Expert (readability, hierarchy, mobile rendering)
- Competitive Positioning Expert (differentiation, unique value)
Email (onboarding, newsletter, campaign)
- Subject Line Specialist (open rate, curiosity, relevance)
- Email Deliverability Expert (spam signals, formatting, length)
- Copywriter (flow, tone, CTA placement)
- Audience Empathy Expert (relevance to reader, pain points)
- Retention Strategist (nurture sequence logic, timing)
Social Post (X, LinkedIn)
- Platform Algorithm Expert (format, timing, engagement signals)
- Hook Specialist (first line, scroll-stopping power)
- Audience Growth Expert (shareability, quote-worthiness)
- Authenticity Analyst (does it sound like a person or a brand?)
Article / Blog
- Editor (structure, flow, argument strength)
- SEO Analyst (keyword relevance, search intent match)
- Reader Engagement Expert (hook, readability, value density)
- Subject Matter Expert (accuracy, depth, credibility)
PR / Communications
- Journalist Perspective (would I cover this? is there news value?)
- Headline Analyst (click-worthiness, specificity)
- Data Density Expert (facts vs. fluff ratio)
- Crisis/Risk Reviewer (could this backfire?)
Strategy / Business
- Data Foundation Expert (are claims data-backed?)
- ROI Analyst (is the business case clear?)
- Risk Assessor (what could go wrong?)
- Actionability Expert (can someone execute this tomorrow?)
For domains not listed above, assemble a panel using this principle:
- 3-4 craft experts (how to make good content of this type)
- 2-3 domain experts (the specific market/audience/subject)
- Humanizer (always)
- Brand Voice Match (always)
- Cap at 10, merge overlapping roles
Step 3: Select Scoring Rubric
Choose the rubric based on content type. Each has 4 dimensions worth 0-25 points (total 0-100):
Content Quality (articles, social posts, newsletters)
| Dimension | 0-25 | What it measures |
|---|---|---|
| Hook Power | /25 | Does the opening stop scrolling? Curiosity gap? Tension? |
| Voice Authenticity | /25 | Does this sound like a human with opinions? Or like AI? |
| Value Density | /25 | Information-to-word ratio. Every sentence earns its place? |
| Engagement Potential | /25 | Would someone share, reply, or save this? |
Conversion Quality (landing pages, emails, ads, CTAs)
| Dimension | 0-25 | What it measures |
|---|---|---|
| Headline/Hero | /25 | Clarity, specificity, relevance to target audience |
| Clarity & Friction | /25 | Can someone understand and act in under 10 seconds? |
| Social Proof & Trust | /25 | Credibility signals, testimonials, trust badges |
| CTA Strength | /25 | Clear, compelling, low-friction call to action |
Strategic Quality (strategies, plans, briefs)
| Dimension | 0-25 | What it measures |
|---|---|---|
| Data Foundation | /25 | Claims backed by specific data, not vibes |
| Actionability | /25 | Can someone execute this tomorrow? |
| ROI Clarity | /25 | Is the business case obvious? |
| Risk Assessment | /25 | Are failure modes identified? |
Evaluation Quality (reviews, audits, assessments)
| Dimension | 0-25 | What it measures |
|---|---|---|
| Evidence Quality | /25 | Specific examples, not vague impressions |
| Criteria Relevance | /25 | Are we measuring what matters? |
| Risk Assessment | /25 | Blind spots identified? |
| Actionability | /25 | Clear next steps with priority? |
Step 4: Score (Recursive Loop)
Round structure
For each round, produce:
- Expert scores table:
| Expert | Score | Key feedback (1 line) |
|---|---|---|
| Humanizer (1.5x) | XX/100 | ... |
| Brand Voice Match | XX/100 | ... |
| [Expert 3] | XX/100 | ... |
Rubric scores: the 4-dimension breakdown
Weighted aggregate: average of all expert scores (humanizer counted at 1.5x weight)
Top 3 weaknesses: ranked, with specific fix suggestions
If score < 90: revise the content addressing the top 3 weaknesses, then run the next round
If score >= 90: finalize
Scoring rules
- Scores must be brutally honest. No padding to reach 90.
- First round scores are typically 55-75. That's normal.
- Each round should show measurable improvement on the identified weaknesses.
- Max 3 rounds. If still under 90 after round 3, ship with the best version and note remaining issues.
- The humanizer score gates everything: if humanizer < 80, the content does not ship regardless of aggregate score.
Quality gates
- < 70: Do not ship. Major rewrite needed.
- 70-79: Shippable with caveats. Note what's weak.
- 80-89: Good. Minor polish possible but not blocking.
- 90+: Ship with confidence.
Step 5: Output
---
**Content type**: [type]
**Panel**: [list of expert names]
**Rubric**: [which rubric used]
**Final score**: XX/100 (after N rounds)
### Rubric breakdown
| Dimension | Score |
|-----------|-------|
| ... | /25 |
| ... | /25 |
| ... | /25 |
| ... | /25 |
### Round history
[Round 1: XX/100 -> top issues -> Round 2: XX/100 -> ...]
### Expert verdicts (final round)
| Expert | Score | Verdict |
|--------|-------|---------|
| ... | ... | ... |
---
[The final scored/revised content]
---
### Remaining issues (if score < 90)
- [what's still weak and why]
Step 6: Feedback-to-Source
When scoring content produced by another skill (e.g., mktg:copy-optimize output, writing:xpost draft):
- Generate a 3-5 line improvement brief that the source skill can use
- Format: "To improve: [specific change]. Because: [expert rationale]."
Memory: Learned Patterns
After each scoring session, if the user approves or rejects the final output, note the pattern in ~/.claude/skills/mktg:expert-panel/patterns.md
Format:
## [Pattern Name]
- **Type:** approval | rejection
- **Content types:** [which types]
- **Rule:** [what to always/never do]
- **Date:** [YYYY-MM-DD]
Version publique de ce Skill. 211 lignes. Copie-colle dans ~/.claude/skills/mktg:expert-panel/SKILL.md pour l'installer.
Skills liés
/mktg:cro-auditAudite une page web sur 8 dimensions de conversion : headline, CTA, social proof, urgency, trust, friction, mobile, speed. Produit un rapport scoré avec des fixes prioritisés.
/mktg:trend-scoutScanne Hacker News, Reddit, Google Trends et X pour trouver les tendances pertinentes sur un domaine. Score chaque tendance et suggère des angles de contenu.
/content:planOrchestre la stratégie de contenu. Scanne les articles, les clippings, les tendances et X pour décider quoi publier, quand et pourquoi. Génère des calendriers éditoriaux.