Ads
Paid acquisition strategy, budget allocation, and avoiding common advertising mistakes across platforms
MIT-0 · Free to use, modify, and redistribute. No attribution required.
⭐ 7 · 810 · 1 current installs · 1 all-time installs
byIván@ivangdavila
MIT-0
Security Scan
OpenClaw
Benign
high confidencePurpose & Capability
The name/description ('Ads' / paid acquisition strategy) matches the content of SKILL.md: marketing best-practices across platforms. It does not request unrelated privileges, binaries, or credentials. Note: the skill's source/homepage are unknown which affects provenance but not internal coherence.
Instruction Scope
SKILL.md contains static advice, checklists, and platform-specific patterns. It contains no commands, no instructions to read files, no env var access, and no network endpoints — the instructions stay strictly within the stated marketing/advice scope.
Install Mechanism
No install spec and no code files are present (instruction-only). Nothing is written to disk or downloaded as part of installation — lowest-risk install profile.
Credentials
No environment variables, credentials, or config paths are requested. The skill does not ask for any secrets or unrelated service keys, which is proportionate for an advisory marketing skill.
Persistence & Privilege
always is false and the skill does not request elevated/persistent presence or modify other skills or system settings. Autonomous invocation is allowed by default but that is normal and not problematic here given the skill's read-only content.
Assessment
This skill is a readable set of ad strategy recommendations and appears coherent and low-risk. Consider these practical precautions before using it: 1) the skill's source and homepage are unknown — verify the publisher if provenance matters; 2) it will only provide advice, not perform actions, so do not paste credentials or account tokens into prompts; 3) validate any platform-specific recommendations against current ad platform docs and account policies (ad platforms change rules frequently); 4) when applying recommendations, test with small budgets and proper tracking (do not let automated advice directly control ad accounts or billing). Overall the skill is safe to install but treat it as advisory content, not a trusted automation for account-level operations.Like a lobster shell, security has layers — review code before you run it.
Current versionv1.0.0
Download ziplatest
License
MIT-0
Free to use, modify, and redistribute. No attribution required.
SKILL.md
Budget Mistakes
- Starting with daily budgets too low to exit learning phase — platforms need 50+ conversions/week per ad set to optimize properly
- Spreading budget across too many campaigns early — concentrate spend to gather statistically significant data faster
- Killing ads before statistical significance — minimum 100 clicks or 1000 impressions before judging creative performance
- No contingency for scaling — reserve 20-30% of budget for doubling down on winners mid-month
- Treating ad spend as expense, not investment — track payback period, not just immediate ROAS
Metric Traps
- Optimizing for CTR instead of conversion — high CTR with low conversion = curiosity clicks that waste budget
- Trusting platform-reported conversions — attribution windows vary (7-day click, 1-day view), always cross-reference with actual revenue
- Ignoring frequency — above 3-4 frequency per week, performance degrades and audience burns out
- CPA tunnel vision — a $50 CPA is better than $30 CPA if LTV is 3x higher for the $50 cohort
- Vanity reach metrics — 1M impressions mean nothing if 0 target customers saw the ad
Creative Rules
- One variable per test — changing image AND copy simultaneously teaches nothing about what works
- Winning ads fatigue in 2-4 weeks — have next creative batch ready before performance drops
- Static images often outperform video on cost-per-conversion — test both, don't assume video is better
- Headlines matter more than body copy — 80% of viewers read only the headline
- User-generated content style outperforms polished brand creative in most direct response contexts
Audience Strategy
- Broad targeting often wins at scale — platform algorithms find converters better than manual interest stacking
- Lookalike audiences need minimum 1000 source users — smaller seeds create unstable lookalikes
- Retargeting pools need 7-14 day recency caps — beyond that, intent has faded
- Exclude converters from prospecting campaigns — paying to show ads to existing customers wastes budget
- Test 1% vs 3% vs 5% lookalikes — tighter isn't always better, depends on market size
Platform-Specific Patterns
- Meta: Learning phase resets with significant edits — avoid editing during first 50 conversions
- Google: Search intent beats display reach for direct response — display is for awareness, search is for capture
- TikTok: First 3 seconds determine everything — hook must be instant, no slow brand intros
- LinkedIn: CPMs are 5-10x higher — only viable for high-LTV B2B where one customer justifies $200+ CPA
- YouTube: Skippable ads teach you what hooks work — if they don't skip, your hook is strong
Scaling Pitfalls
- Increasing budget more than 20-30% per day destabilizes campaigns — gradual scaling preserves algorithm learning
- Duplicating winning ad sets fragments the audience and causes self-competition
- Scaling spend without scaling creative — same ads to larger audience = faster fatigue
- Ignoring incrementality — some conversions would have happened organically, true ROAS is lower than reported
- Geographic expansion without localization — same ad in new market often fails
Landing Page Impact
- Ads are only half the equation — a 2x better landing page beats 2x more ad spend
- Message match: ad promise must appear above the fold on landing page — disconnect kills conversion
- Page load time over 3 seconds loses 50%+ of paid clicks — optimize speed before scaling spend
- One landing page per audience segment — generic pages convert worse than specific ones
- Track micro-conversions (scroll depth, time on page) when sample size is too small for macro-conversions
Attribution Reality
- Last-click attribution undervalues awareness campaigns — multi-touch attribution or holdout tests reveal true impact
- iOS 14.5+ broke tracking for ~40% of users — model conversions, don't rely on pixel data alone
- Offline conversions (calls, in-store) need manual import or integration — otherwise CPA looks inflated
- View-through conversions are real but overvalued by platforms — weight click-through higher
- 7-day attribution windows miss longer B2B sales cycles — extend windows or use CRM-based attribution
Testing Framework
- Always run one control ad — without baseline, you don't know if new creative is better or platform just performed differently
- Minimum 2 weeks per test — weekday/weekend patterns affect results
- Document every test with hypothesis, result, and learning — institutional memory prevents repeat mistakes
- Test audiences before creatives — wrong audience can't be saved by good creative
- Negative results are valuable — knowing what doesn't work prevents future waste
Files
1 totalSelect a file
Select a file to preview.
Comments
Loading comments…
