--- name: ad-campaign-analyzer description: > Analyze ad campaign performance data (Google, Meta, LinkedIn) to identify what's working, what's wasting budget, and specific cut/scale/test recommendations. Runs statistical analysis, funnel diagnostics, and multi-channel budget reallocation with specific dollar-amount shift recommendations and scenario modeling. tags: [ads] --- # Ad Campaign Analyzer Take raw campaign performance data and turn it into clear decisions. This skill doesn't just summarize metrics — it diagnoses problems, identifies winners, checks statistical significance, and tells you exactly what to cut, scale, and test next. Then it goes further: it compares channels on equal terms, finds where you're over-spending vs under-spending relative to results, and produces a concrete budget reallocation plan. **Core principle:** Most startup founders check their ad dashboard, see a ROAS number, and either panic or celebrate. This skill gives you the nuanced analysis a paid media specialist would: what's actually significant, what's noise, and where your next dollar should go. It also solves the allocation problem — most startups either spread budget too thin across channels (no channel gets enough to learn) or dump everything into one channel (missing cheaper opportunities elsewhere). ## When to Use - "Analyze my Google Ads performance" - "Which ads should I kill?" - "Is this campaign working?" - "Where am I wasting ad spend?" - "Optimize my Meta Ads" - "How should I split my ad budget?" - "Should I spend more on Google or Meta?" - "Reallocate my ad spend across channels" - "Where am I getting the best return?" - "I have $X/month for ads — how should I distribute it?" ## Phase 0: Intake 1. **Campaign data** — One of: - CSV export from Google Ads / Meta Ads Manager / LinkedIn Campaign Manager - Pasted performance table - Screenshots of dashboard (we'll extract the data) 2. **Platform(s)** — Google / Meta / LinkedIn / All 3. **Time period** — What date range does this cover? 4. **Monthly budget** — Total ad spend in this period 5. **Primary goal** — What conversion are you optimizing for? (Demos / Trials / Purchases / Leads) 6. **Target metrics** — Do you have target CPA or ROAS? (If not, we'll benchmark) 7. **Any known changes?** — Did you change creative, budget, or targeting during this period? 8. **Channels currently running** — Google Ads, Meta Ads, LinkedIn Ads, Twitter/X Ads, TikTok Ads, other 9. **Funnel data** (if available): - Lead → MQL rate - MQL → SQL rate - SQL → Close rate - Average deal size 10. **Channels you're considering but haven't tried** — Want to test new channels? 11. **Constraints** — Minimum spend on any channel? Platform you must stay on? ## Phase 1: Data Ingestion & Normalization ### Accepted Data Formats | Source | Key Columns Expected | |--------|---------------------| | **Google Ads** | Campaign, Ad Group, Keyword, Impressions, Clicks, CTR, CPC, Conversions, Conv Rate, Cost, Conv Value | | **Meta Ads** | Campaign, Ad Set, Ad, Impressions, Reach, Clicks, CTR, CPC, Conversions, Cost Per Result, Amount Spent, ROAS | | **LinkedIn Ads** | Campaign, Impressions, Clicks, CTR, CPC, Conversions, Cost, Leads | Normalize all data into a standard analysis format: | Dimension | Impressions | Clicks | CTR | CPC | Conversions | Conv Rate | CPA | Spend | Revenue/Value | |-----------|------------|--------|-----|-----|-------------|----------|-----|-------|--------------| ### Multi-Channel Normalization When data spans multiple channels, also produce a channel-level rollup: | Channel | Monthly Spend | Impressions | Clicks | CTR | CPC | Conversions | Conv Rate | CPA | ROAS | CAC* | |---------|-------------|------------|--------|-----|-----|-------------|----------|-----|------|------| | Google Search | $[X] | [N] | [N] | [X%] | $[X] | [N] | [X%] | $[X] | [X] | $[X] | | Google Display | ... | | | | | | | | | | | Meta (FB/IG) | ... | | | | | | | | | | | LinkedIn | ... | | | | | | | | | | | [Other] | ... | | | | | | | | | | | **Total** | $[X] | | | | | [N] | | $[X] avg | [X] avg | $[X] avg | *CAC = Full customer acquisition cost if funnel data provided (CPA × close-rate adjustment) ### Funnel-Adjusted CAC (If Funnel Data Available) ``` Channel CAC = CPA ÷ (MQL rate × SQL rate × Close rate) ``` This reveals which channels produce leads that actually close, not just convert. ## Phase 2: Performance Diagnostics ### 2A: Campaign-Level Health Check For each campaign: | Metric | Value | Benchmark | Status | |--------|-------|-----------|--------| | CTR | [X%] | [Industry avg] | [Good/Okay/Poor] | | CPC | $[X] | [Category avg] | [Good/Okay/Poor] | | Conv Rate | [X%] | [Benchmark] | [Good/Okay/Poor] | | CPA | $[X] | [Target or benchmark] | [Good/Okay/Poor] | | ROAS | [X] | [Target or benchmark] | [Good/Okay/Poor] | | Impression Share | [X%] | [>60% ideal] | [Good/Okay/Poor] | ### 2B: Budget Waste Detection Identify spend that produced no or negative return: | Waste Type | Signal | Action | |-----------|--------|--------| | **Zero-conversion keywords/ads** | Spend > $[X] with 0 conversions | Pause or add negatives | | **High CPA outliers** | CPA > 3x target | Pause or restructure | | **Low CTR ads** | CTR < 50% of campaign average | Replace creative | | **Broad match bleed** | Search terms report showing irrelevant clicks | Add negative keywords | | **Audience overlap** | Same users hit by multiple campaigns | Exclude audiences | | **Dayparting waste** | Conversions cluster at certain hours; spend is 24/7 | Set ad schedule | ### 2C: Winner Identification Find what's actually working: | Winner Type | Signal | Action | |------------|--------|--------| | **Top-performing keywords** | Lowest CPA, highest conv rate | Increase bid, add variants | | **Winning ads** | Highest CTR + conv rate combo | Scale spend, clone for other groups | | **Best audiences** | Lowest CPA segment | Increase budget allocation | | **Best times** | Peak conversion hours/days | Concentrate budget | ### 2D: Statistical Significance Check For any A/B test (ad variants, audiences, landing pages): ``` Test: [Variant A] vs [Variant B] Metric: [Conv Rate / CTR / CPA] Variant A: [X%] (n=[sample_size]) Variant B: [Y%] (n=[sample_size]) Confidence level: [X%] Verdict: [Statistically significant / Not enough data / Too close to call] Recommended action: [Pick winner / Continue test / Increase budget to reach significance] ``` Minimum sample: 100 clicks per variant for CTR tests, 30 conversions per variant for CPA tests. ## Phase 3: Funnel Analysis ### Click → Conversion Path ``` Impressions: [N] (100%) ↓ CTR: [X%] Clicks: [N] ([X%] of impressions) ↓ Landing page → Conversion: [X%] Conversions: [N] ([X%] of clicks) ↓ Conversion → Revenue: $[X] avg Revenue: $[N] ``` ### Funnel Drop-Off Diagnosis | Drop-Off Point | Rate | Benchmark | Likely Cause | Fix | |----------------|------|-----------|-------------|-----| | Impression → Click | [CTR%] | [Benchmark] | [Ad relevance / targeting] | [Copy/targeting change] | | Click → Conversion | [Conv%] | [Benchmark] | [Landing page / offer / audience mismatch] | [LP optimization] | | Conversion → Revenue | [Close%] | [Benchmark] | [Lead quality / sales process] | [Qualification criteria] | ## Phase 4: Budget Reallocation When data spans multiple channels, perform cross-channel budget optimization. ### 4A: Channel Efficiency Ranking | Rank | Channel | CPA | Funnel-Adj CAC | Share of Spend | Share of Conversions | Efficiency Index | |------|---------|-----|---------------|----------------|---------------------|-----------------| | 1 | [Channel] | $[X] | $[X] | [X%] | [X%] | [Conv share ÷ Spend share] | **Efficiency Index:** - **> 1.0** = Under-invested (getting more than its share of conversions) - **= 1.0** = Proportional (fair share) - **< 1.0** = Over-invested (getting less than its share) ### 4B: Marginal Return Analysis For each channel, estimate if additional spend would yield proportional returns: | Channel | Current CPA | Impression Share / Saturation Signal | Marginal Return Estimate | |---------|-------------|-------------------------------------|------------------------| | Google Search | $[X] | [X%] impression share — room to grow | Likely positive | | Meta | $[X] | Frequency [X] — audience may be saturated | Diminishing | | LinkedIn | $[X] | Low volume — limited targeting pool | Ceiling soon | ### 4C: Funnel Stage Coverage | Funnel Stage | Channels Covering It | Current Spend | Gap? | |-------------|---------------------|--------------|------| | **Awareness** (top) | [Meta Display, YouTube] | $[X] | [Yes/No] | | **Consideration** (mid) | [Google Search, Meta retargeting] | $[X] | [Yes/No] | | **Decision** (bottom) | [Google Brand, Google Search] | $[X] | [Yes/No] | | **Retargeting** | [Meta, Google Display] | $[X] | [Yes/No] | ### 4D: Budget Shift Recommendations | Channel | Current Spend | Recommended Spend | Change | Reasoning | |---------|-------------|------------------|--------|-----------| | Google Search | $[X] | $[Y] | +$[Z] | [Lowest CPA, room to scale] | | Meta | $[X] | $[Y] | -$[Z] | [Audience saturation, frequency too high] | | LinkedIn | $[X] | $[Y] | $0 | [Maintain — niche but valuable] | | [New channel] | $0 | $[Y] | +$[Y] | [Test budget — competitors succeeding here] | | **Total** | $[X] | $[X] | $0 | Budget-neutral reallocation | ### 4E: Scenario Modeling **Scenario 1: Conservative shift (+/- 20%)** - Expected conversions: [N] (currently [N]) = [X%] improvement - Expected blended CPA: $[X] (currently $[X]) - Risk: Low **Scenario 2: Aggressive shift (+/- 40%)** - Expected conversions: [N] = [X%] improvement - Expected blended CPA: $[X] - Risk: Medium — less data on scaled channels **Scenario 3: Budget increase to $[Y]/mo** - Recommended allocation: [table] - Expected conversions: [N] - New channels to test: [list] ## Phase 5: Output Format ```markdown # Ad Campaign Analysis — [Product/Client] — [DATE] Period: [Date range] Total spend: $[X] Platform(s): [Google / Meta / LinkedIn] Primary goal: [Conversions / Revenue / Leads] --- ## Executive Summary [3-5 sentences: Overall performance verdict, biggest win, biggest problem, top recommendation including any reallocation moves] --- ## Performance Dashboard | Campaign | Spend | Impressions | Clicks | CTR | CPC | Conversions | CPA | ROAS | Verdict | |----------|-------|------------|--------|-----|-----|-------------|-----|------|---------| | [Name] | $[X] | [N] | [N] | [X%] | $[X] | [N] | $[X] | [X] | [Scale/Optimize/Pause] | --- ## Budget Waste Report **Total estimated waste: $[X] ([X%] of total spend)** ### Wasted on zero-conversion items: $[X] [List of keywords/ads/audiences with spend but no conversions] ### Wasted on high-CPA items: $[X] [List of items with CPA > 3x target] ### Recommended saves: $[X]/month [Specific items to pause] --- ## Winners to Scale ### Top Keywords/Audiences | Item | CPA | Conv Rate | Current Spend | Recommended Spend | |------|-----|----------|--------------|-------------------| ### Top Ads | Ad | CTR | Conv Rate | Why It Works | |----|-----|----------|-------------| --- ## A/B Test Results ### [Test Name] - Variant A: [Metric] (n=[N]) - Variant B: [Metric] (n=[N]) - Confidence: [X%] - **Verdict:** [Winner / Continue / Inconclusive] --- ## Budget Reallocation ### Current vs Recommended Allocation | Channel | Current | Recommended | Change | Why | |---------|---------|------------|--------|-----| | [Channel] | $[X] | $[Y] | [+/-$Z] | [1-line reason] | **Projected impact:** - Conversions: [N] → [N] (+[X%]) - Blended CPA: $[X] → $[Y] (-[X%]) ### Funnel Stage Coverage [Coverage map with gaps identified] ### New Channel Recommendations #### [Channel Name] - **Why test:** [Reasoning] - **Recommended test budget:** $[X]/mo for [X weeks] - **Success criteria:** CPA < $[X] - **Competitors using it:** [Yes/No — who] --- ## Action Plan ### Immediate (This Week) - [ ] **Pause:** [Specific items — keywords, ads, audiences] - [ ] **Scale:** [Specific items — increase budget/bids] - [ ] **Add negatives:** [Specific keywords from search terms] - [ ] **Reallocate:** [Specific dollar shifts between channels] ### This Month - [ ] **Test:** [New ad angles / audiences / landing pages] - [ ] **Restructure:** [Ad groups that need splitting or merging] - [ ] **Optimize:** [Bid strategy changes] - [ ] **Monitor reallocation:** Track CPA shifts on scaled channels, watch for diminishing returns ### Next Month - [ ] **Expand:** [New campaigns / channels to test] - [ ] **Re-evaluate:** [Run this analysis again with new data, adjust allocations based on actual results] ``` Save to `campaign-analysis-[YYYY-MM-DD].md` in the current working directory (or user-specified path). ## Cost | Component | Cost | |-----------|------| | Data analysis | Free (LLM reasoning) | | Statistical calculations | Free | | **Total** | **Free** | ## Tools Required - No external tools needed — pure reasoning skill - User provides campaign data as CSV, paste, or screenshot ## Trigger Phrases - "Analyze my ad campaign performance" - "Which ads should I pause?" - "Where am I wasting ad budget?" - "Is my Google Ads campaign working?" - "Optimize my Meta Ads spend" - "How should I allocate my ad budget?" - "Should I spend more on Google or Meta?" - "Reallocate my ad spend" - "Where am I getting the best ROAS?" - "Optimize my multi-channel ad budget"