You built something people use. Now you want more people to find it—without accidentally burning $5,000 on a single weekend or turning into a full-time media buyer.
Most engineers hit the same wall: open Facebook Ads Manager or Apple Search Ads, see 47 toggles and acronyms (CPM? oCPM? CBO?), panic, and either overspend or give up. The alternative—hiring an agency—often means $3k/month retainers before a single install.
There’s a middle path: structured micro-tests that give you real signals in 5–7 days with a few hundred dollars. No marketing degree required. Just guardrails, a tight brief, and basic spreadsheet logic.
Why most first UA tests fail (and waste money)
Three common mistakes:
- No success threshold defined upfront. You spend $800, get 150 installs, then wonder “is that… good?” without a payback model.
- Too many variables at once. Five ad networks, ten creatives, four audiences = impossible to learn what worked.
- Missing conversion events. You optimize for installs but never fire a “trial_start” or “purchase” event, so the algorithm learns nothing about quality.
Result: you either pause everything in fear or let campaigns run on autopilot until the credit card alert arrives.

The MVP stack: one network, three creatives, one audience
Start with the smallest falsifiable test:
- One network: Apple Search Ads (if iOS) or Meta (if your app has broad appeal). Pick the one where your likely users already scroll.
- Three creatives: One feature demo, one testimonial/review highlight, one problem/solution hook. Keep them under 15 seconds or six static frames.
- One audience: Broad targeting (age 25–45, interests related to your category) or keyword match (ASA). Avoid hyper-narrow segments until you have baseline data.
Why this constraint? You need at least 50–100 conversions per variant to see signal. Splitting $500 across ten ad sets gives you noise, not learning.
Guardrails: the safety rails that let you move fast
Before you launch, set these four limits in the platform:
| Guardrail | What it does | Example setting |
|---|---|---|
| Daily budget cap | Maximum you can spend in 24 hours | $75/day for a $500 test |
| Bid cap (CPI or CPA) | Max price you’ll pay per install or action | $4.00 CPI if your LTV is $12+ |
| Auto-pause rule | Stop ad set if CPI exceeds threshold after X installs | Pause if CPI > $6 after 30 installs |
| Conversion window | How long to wait for events (trial, purchase) before judging quality | 7-day click, 1-day view |
These settings prevent the “wake up to $900 gone” scenario. You’re buying information, not scale—yet.

What to test (and what to ignore for now)
High-signal, low-complexity tests for week one:
- Creative angle: Does “Save 2 hours a week” outperform “Built by designers, for designers”? Run both and compare install-to-trial rate.
- Store asset A/B: Test two icon variants or first screenshot frames via App Store experiments (free, built-in).
- Paywall timing: Trial-gate on launch vs. after one task completion. Track trial start rate and D1 retention.
Ignore for now:
- Lookalike audiences (need 1,000+ converters first)
- Retargeting (wait until you have 10k+ app opens)
- Advanced attribution (MMP setup can wait until you’re spending $2k+/month)
You’re looking for one repeatable winner—a creative + audience combo that hits your target CPI with acceptable trial conversion. Once you find it, that’s when you scale or bring in help.
Logging: the four events that matter
Ad platforms optimize toward events you send them. If you only fire “install,” the algorithm delivers installers—not necessarily users who pay. Wire up these four:
- Install (automatic on iOS/Android)
- First open / onboarding complete (shows app isn’t crashing)
- Trial start (the user saw value and opted in)
- Purchase (actual revenue event)
Use the native SDKs (StoreKit for Apple Search Ads, Meta SDK for Facebook/Instagram) or a simple webhook to your analytics. You don’t need a full MMP like Adjust or AppsFlyer yet—those cost $100–500/month and add complexity.
Why this matters: After 20–30 purchases, Meta or Google can start optimizing for “purchase” instead of “install,” which drops your effective CAC by 20–40%. But only if you’re sending the event.

Worked example: $500 test → next-step logic
You run a 7-day test on Meta with $75/day budget, targeting a broad audience (25–50, interested in “productivity apps”). Here’s what you get:
| Metric | Result |
|---|---|
| Spend | $525 (slight overage on day 6) |
| Impressions | 68,000 |
| Clicks | 1,200 |
| Installs | 210 |
| CPI | $2.50 |
| Trial starts | 13 (6.2% of installs) |
| Purchases (D7) | 2 ($9.99 each) |
Quick math:
- Revenue so far: $19.98
- Payback at D7: 3.8% ($19.98 ÷ $525)
- Projected LTV (if your trial → paid rate is 25% and annual retention is 60%): ~$15 per install over 12 months
- Break-even CPI: $15 (acceptable if you have 6+ month payback tolerance)
Decision tree:
- If CPI < $3 and trial rate > 5%: Increase budget to $150/day, add two more creatives, test a second audience.
- If CPI = $2–4 and trial rate = 3–5%: Optimize onboarding or paywall timing before scaling. The unit economics almost work.
- If CPI > $5 or trial rate < 2%: Pause. Either the creative is off, the product-market fit isn’t tight, or the audience is wrong. Revisit positioning.
This test cost you $525 and gave you a falsifiable answer in one week. That’s cheaper than most “strategy consults” and infinitely more useful than guessing.
Template: your one-page test plan
Before you launch, fill this out. It forces clarity and prevents mid-flight panic:
| Field | Your answer |
|---|---|
| Goal | Validate CPI < $4 with 5%+ trial rate |
| Budget | $500 over 7 days |
| Network | Meta (Facebook + Instagram feed) |
| Audience | 25–45, US, interests: productivity, notion, asana |
| Creatives | 1) Feature demo (15s), 2) User testimonial (static), 3) Problem/solution hook (10s) |
| Bid cap | $4.50 CPI |
| Daily cap | $75 |
| Auto-pause rule | Stop ad set if CPI > $6 after 40 installs |
| Events tracked | Install, first_open, trial_start, purchase |
| Success criteria | CPI < $4 and trial rate > 5% |
| Next step if success | 2x budget, add 2 creatives, test lookalike audience |
| Next step if fail | Pause, revisit onboarding flow or creative angle |

When to do it yourself vs. bring in help
DIY makes sense when:
- Your monthly budget is under $2,000
- You want to learn the mechanics (useful for future products)
- You have 3–5 hours/week to monitor and tweak
- Your app is early—still iterating core features
Consider a contractor or freelancer ($500–1,500/month) if:
- You’ve validated CPI < payback threshold and want to scale to $5k+/month
- You need creative production (video editing, static design) on a regular cadence
- You’d rather spend the time on product than dashboards
Consider an agency ($3k–10k/month retainer) if:
- You’re spending $15k+/month and need multi-channel orchestration (Meta, Google, ASA, TikTok)
- You want strategic planning, not just execution
- You have budget for a 3–6 month commitment
Consider a publishing partner (rev-share, no upfront cost) if:
- You’ve proven product-market fit (D7 retention > 15%, clear monetization)
- You want someone to fund and operate UA/ASO/creative testing while you focus on product
- You prefer aligned incentives (they win when you win) over fixed fees
Most builders start DIY, hit a ceiling around $2k/month spend (too manual to scale, not enough volume to justify an agency), and either plateau or look for a partner who can take over growth ops.
Common mistakes (and how to dodge them)
- “I’ll just boost this post.” Boosted posts on Meta rarely optimize for app installs—they optimize for engagement. Use Campaign Manager with an “App Installs” objective.
- “I’ll target super narrow (e.g., ‘SaaS founders in SF’).” You’ll get 12 impressions. Start broad, let the algorithm find patterns.
- “I’ll test ten creatives at once.” You need 50+ conversions per variant to see significance. Three is the max for a $500 test.
- “I don’t need to track trials, just installs.” Then you’ll optimize for tire-kickers, not buyers. Always ladder up to a revenue event.

The spreadsheet: track your test in real time
You don’t need fancy BI tools. A simple CSV or Google Sheet with these columns will do:
| Date | Spend | Impressions | Clicks | Installs | CPI | Trial starts | Trial % | Purchases | Revenue |
|---|---|---|---|---|---|---|---|---|---|
| Oct 7 | $72 | 9,800 | 165 | 28 | $2.57 | 2 | 7.1% | 0 | $0 |
| Oct 8 | $75 | 10,100 | 178 | 31 | $2.42 | 2 | 6.5% | 1 | $9.99 |
| Oct 9 | $78 | 10,500 | 182 | 33 | $2.36 | 1 | 3.0% | 0 | $0 |
| Total | $525 | 68,000 | 1,200 | 210 | $2.50 | 13 | 6.2% | 2 | $19.98 |
Update it daily (takes 2 minutes). By day 4, you’ll see if the test is on track or needs a mid-flight tweak (pause underperforming ad set, shift budget to winner).
What happens after the test
You now have one of three outcomes:
- Clear win: CPI under target, trial rate above 5%, payback horizon looks reasonable. → Scale to $150–300/day, add more creatives, test a second audience or network.
- Marginal: CPI is acceptable but trial rate is weak, or vice versa. → Fix the weak link (onboarding flow, paywall copy, creative hook) and re-test in 2 weeks.
- No signal: CPI way over target, trial rate under 2%, or both. → Pause. Revisit your positioning, ideal user, or core feature set before spending more.
The worst outcome is no decision—letting the test run without guardrails, burning $2k, and still not knowing if your unit economics work. A structured $500 test prevents that.
When a test works: what scaling looks like
Once you’ve found a repeatable winner (stable CPI, acceptable payback), you have three paths:
- Self-funded scale: Increase budget 2x every week until performance degrades or you hit cash flow limits. Requires close monitoring and regular creative refreshes.
- Hire execution help: Bring in a freelance media buyer or performance marketer ($1–2k/month) to manage the day-to-day while you focus on product.
- Partner with a publisher: If you’d rather keep building and let someone else fund + operate growth, a rev-share publishing partner (like TorApps) can take over UA, ASO, creative testing, and reporting while you maintain product velocity.
The key insight: you don’t need to become a marketer to validate that marketing works. You just need a tight test, clear metrics, and the discipline to pause what doesn’t work.

Next steps: start your first test this week
Here’s your checklist to launch in the next 5 days:
- [ ] Pick one network (Apple Search Ads or Meta)
- [ ] Create three creatives (one feature demo, one testimonial, one problem/solution)
- [ ] Set up conversion events (install, trial_start, purchase)
- [ ] Define success criteria (target CPI + trial rate threshold)
- [ ] Set guardrails (daily cap, bid cap, auto-pause rule)
- [ ] Launch with $75/day budget
- [ ] Log results daily in the tracking sheet
- [ ] Make a go/no-go decision on day 7
That’s it. No certification courses, no $5k agency onboarding, no six-month commitment. Just a structured week of learning.
If you want a partner to fund and scale what’s working
Once you’ve validated that your unit economics work—CPI under your payback threshold, trial-to-paid rate above 20%, D7 retention solid—the bottleneck often shifts from “does this work?” to “how do we scale this without burning all our time on dashboards?”
That’s where a publishing partner makes sense. Instead of paying upfront retainers or giving away equity, you align incentives: they fund and operate UA, ASO, creative testing, and analytics; you keep building product. Revenue is shared based on what actually gets generated.
How it typically works:
- You keep IP and code ownership
- The partner gets publishing rights and handles all growth ops
- You stay focused on shipping features and fixing bugs
- Both sides win when the app wins—no misaligned incentives
If you’ve already run a successful test (or have a live app with decent retention), we’d be happy to review your numbers and see if there’s a fit. No pitch decks, no multi-month diligence—just a quick look at your metrics and roadmap.



