Killian Brief
April 29, 2026 · Nightly Run · 5 Bets Shortlisted
Bets shortlisted
5
Avg judge score
74/100
Run cost
$3.79
Bet #1

AI screenshots that match your actual app

judge 76/100edge 2.0/10ai native

Every app release, a solo dev burns half a day in Figma faking screenshots — or pays a designer $250 and waits three days. The result still looks like a template. Meanwhile the top 100 apps on the store have screenshots so polished they convert 2x. That gap is money on the table every single ship cycle.

There are roughly 50-100k devs shipping monthly updates to iOS/Android. Capture 1.5% at $39/mo blended ARPU and that's ~$700k ARR — not a unicorn, but a real lifestyle SaaS with a clear $1M+ ceiling if we land small studios managing multiple apps.

Incumbents (AppLaunchpad, Shotbot, Appscreens) are template factories — the dev still does the design work. Our wedge: ingest the user's raw screens, extract their actual UI palette/typography/tone, and output on-brand assets across all device/locale combos. Why now: multimodal models got cheap and good enough at style transfer in the last 12 months to make this economically viable per generation.

Honest on edges: Lisandro's Apple background helps with App Store mechanics but isn't a moat. Defensibility is the real risk — Canva can clone this in 60 days. So we don't build a fortress, we build a cashflow sprint.

The bet: 14 days, ~$2k in API + landing page + Vercel. Kill if <15 signups, <25% complete a generation, or <3 paid conversions by day 30. Small, reversible, fast signal.

Let's ship it before Canva wakes up.

The detail behind the pitch
Problem
App developers lose 0.5 days per release creating App Store screenshots, either hacking rough versions in design tools or paying designers $200-300 and waiting 3 days.
Proposed solution
AI tool that auto-generates marketing screenshots in correct format/devices by extracting visual style from top apps and restyling user's raw captures.
Target market
Solo and small-team app developers shipping to iOS/Android (estimated 50k-100k active monthly who release updates); currently paying designers or losing time.
First test
Launch free tier (first credit free), track: (1) signups in 14d, (2) % who upload screens, (3) % who regenerate/pay. Kill if <20 signups or <30% conversion to paid action.
Kill criteria
<15 signups by day 14 OR <25% of signups complete a full generation (upload + receive output) by day 14 OR <3 paid conversions ($29+) by day 30 OR avg user quality rating of generated screenshots <3.5/5 (collected via mandatory post-generation prompt) by day 30 → kill or full pivot
Competitive landscape
Incumbents: AppLaunchpad, Shotbot, Previewed, ScreenshotOne, MockuPhone, Rottenwood (now Appscreens.io), Canva (App Screenshot templates), Zapp (by Gummicube) Pricing: $0 (limited free tier) to $29-$79/mo for most SaaS tools; one-off designer cost $200-$300 Saturation: medium Wedge: AI style-extraction that analyzes the user's actual app UI (colors, typography, tone) and auto-generates on-brand, device-correct screenshots without any manual design work — closing the gap between generic template tools and bespoke designer output. User complaints: Templates look generic and don't match the visual style of the actual app; Manual re-sizing across 6+ device/locale combinations is tedious even with tools; Tools require significant design skill to get polished results — not truly automated; No tool extracts or mirrors the real in-app visual language; style must be rebuilt from scratch; Localization (different screenshot text per store locale) is painful and mostly unsupported; Free tiers watermark or severely limit exports, forcing paid plans for basic use; Output often looks templated/amateurish compared to top-charting app screenshots Notes: The space has several template-based SaaS players (AppLaunchpad, Shotbot, Appscreens.io) with commodity $20-$80/mo pricing, but none use AI to infer visual style from the app itself — they all require the developer to design within a template. The real pain is not format/sizing (solved) but brand-coherent, conversion-optimized output without design effort. An AI-first approach that ingests raw screenshots and outputs polished, on-brand store assets has a credible wedge, though defensibility is low once Canva or a larger ASO platform adds the same AI layer. The 50k-100k TAM is plausible but skews toward price-sensitive solo devs, so monetization should target small studios and agencies managing multiple apps.
Skeptic + judge rationale
Death modes: - AI style-extraction outputs look polished in demos but fail on real-world apps with complex UIs (dark mode, custom fonts, cluttered screens) — first 50 users get mediocre results, post brutal honest reviews on indie dev forums (IndieHackers, r/iOSProgramming), word spreads that 'it's just another template tool with an AI wrapper', organic growth flatlines by day 30 - Free tier's single credit is enough: developer uses 1 free generation per app release (which happens every 4-8 weeks), has zero urgency to upgrade, MRR stays under $200 at day 60 because the purchase trigger (a release) is infrequent and the free credit resets the clock — churn kills LTV before CAC is ever recovered - The 3 largest incumbents (AppLaunchpad, Appscreens.io, Canva) each ship an 'AI auto-style' feature within 60 days of launch — they have the distribution (100k+ existing users), the device-frame libraries, and the API budget to clone the wedge instantly, collapsing differentiation before the startup has enough paying customers to fund iteration # Judge rationale (score=76.0) Wins on operational shape: pure software, low capital, Vercel-deployable, minimal human-in-loop once shipped, and a fast path to paying customers via a clear pain ($200-300 designer alternative). Loses on defensibility (1/5) — Canva/AppLaunchpad can clone the AI layer in 60 days with superior distribution. ARPU is mid ($29-79/mo skewing solo-dev price-sensitive) and recurring revenue is shaky because release cadence is every 4-8 weeks, weakening subscription stickiness toward transactional behavior. Market is real but crowded; the wedge is credible but narrow.
Reply "approve #1" on Telegram to ship this bet.
Bet #2

Cash-flow GPS for POD founders quitting jobs

judge 78/100edge 1.5/10b2b saas

Every month, thousands of print-on-demand sellers stare at a Printify dashboard that shows what already happened and try to guess whether they can quit their day job. They're flying blind on the one decision that actually matters: when to push ad spend, when to hold, when to cut. Facebook groups and gut feel are the current 'tools.' That's broken, and it's the difference between freedom and a humiliating return to a cubicle.

The pool is real but narrower than the headline: ~50-100k US POD sellers, but the honest willing-to-pay segment is sellers already doing $1-3k/month who can taste $4k. Call it 5k reachable founders × 3% capture × $50 ARPU = ~$90k ARR year one — a probe, not a unicorn.

The wedge: every cash-flow tool (Pulse, Causal, LivePlan) assumes inventory and demands setup; every POD dashboard is backward-looking. Nobody bakes in platform fee stacks plus prescriptive 'scale/hold/cut' triggers tied to margin thresholds. That gap is real, but copyable by Printify in a quarter — so speed matters.

Why now: honestly, nothing dramatic changed. The opening is just persistent neglect by incumbents.

Why us: weak. This isn't in your manufacturing/Amazon wheelhouse, and I won't pretend otherwise.

The path is cheap and brutal: free Sheet + scaling 1-pager posted to r/Etsysellers and r/shopify, ~$0 capital, 14 days. Kill if <25 signups in 7 days, <3 Stripe reservation clicks in 14, or zero paid conversions in 30. The Sheet-cannibalizes-SaaS risk is real — that's exactly what the paywall test exposes.

Let's spend two weeks and a weekend to find out if anyone actually swipes a card.

The detail behind the pitch
Problem
Print-on-demand shop owners quitting day jobs need to hit $4k/month in 6 months but lack playbooks for when to scale ad spend, inventory, and operational capacity.
Proposed solution
A cash-flow forecasting dashboard that inputs current margins ($20-40/sale), converts them to monthly targets, and recommends weekly scaling rules (when to increase ad budget, add SKUs, hire).
Target market
POD/print shop founders considering full-time transition; ~50-100k in US; willing to pay $30-100/month for risk reduction.
First test
Build a simple Google Sheet template + 1-pager on scaling rules. Post in r/Etsysellers and r/shopify, capture email signups, send 3 follow-up surveys over 14 days asking if they'd pay for the SaaS version.
Kill criteria
<25 email signups from Reddit posts within 7 days of posting AND <3 respondents click a 'reserve my spot for $30/month' Stripe checkout link (not just say yes in a survey) within 14 days AND 0 paid conversions within 30 days of presenting a real paywall → kill
Competitive landscape
Incumbents: Printify Pop-Up Store analytics, Printful Dashboard, Shopify Financial Reports, LivePlan, Pulse (cash flow app), Causal, Fathom HQ Pricing: $19-$75/seat/mo (general cash flow tools); POD-native analytics typically free within platform Saturation: low Wedge: POD-specific cash flow logic (zero inventory, per-order margin, platform fee stacks) baked in by default — no setup required — plus prescriptive weekly action triggers (scale/hold/cut) that generic tools never provide. User complaints: Printful/Printify dashboards show revenue but give zero forward-looking cash flow projections; Generic tools like LivePlan require manual entry of POD-specific cost structures (base cost, shipping, platform fees) — no native integration; No tool tells POD sellers *when* to scale ad spend relative to margin thresholds — they rely on guesswork or Facebook group advice; Existing forecasting tools assume inventory holding costs; POD founders have no inventory but need demand-side scaling rules instead; Etsy/Shopify analytics are backward-looking only — no scenario modeling for 'what if I 3x my ad budget next week' Notes: No known direct competitor targets the 'quit my day job in 6 months' POD founder with prescriptive scaling rules. General cash flow tools (Pulse, Causal, Fathom) serve SMBs broadly but require heavy customization for POD economics and offer no actionable 'when to scale' recommendations. The closest adjacent product is a Shopify or Printify built-in dashboard, which is purely descriptive and backward-looking. The $30-100/month price point is well within range — Pulse charges $29-89/mo for far less POD-relevant insight — and the emotional ROI (confidence to quit a job) is a strong willingness-to-pay driver that generic tools completely ignore.
Skeptic + judge rationale
Death modes: - Reddit/community validation is fatally polluted by 'aspirational yes' bias: r/Etsysellers respondents say they'd pay $30+/month in a survey but the actual conversion rate when a payment link appears is <2%, because the target user is pre-revenue or sub-$500/month MRR and $30/month feels material — the 5 'yes I'd pay' survey responses never translate to a single credit card swipe within 30 days of launching the paid product - The Google Sheet template solves the problem well enough that it permanently kills willingness to pay for the SaaS: users download it, customize it once, share it in Facebook POD groups, it goes viral as a free resource, and the 'aha moment' is fully satisfied at $0 — upgrade rate stays at 0% because the template IS the MVP and there's no remaining pain to monetize - The target user — POD founder considering quitting their day job — is structurally in a 3-12 month pre-decision window and will not pay for a forecasting tool until they're already at $2k+/month; the actual addressable market of POD sellers who are (a) close enough to $4k/month to care, (b) not already using an accountant or spreadsheet, and (c) willing to pay right now is fewer than 500 people in the US, producing a ceiling of ~$15k ARR that can't sustain the product before runway expires # Judge rationale (score=78.0) Wins big on capital (Sheet + Reddit posts cost ~$0), pure-software SaaS with low ops burden, and clean recurring revenue at a defensible $30-100/mo price. Loses points on ARPU (mid-tier prosumer pricing), market (skeptic's point about pre-revenue aspirational buyers shrinks the actual willing-to-pay pool to maybe 1-10k), and defensibility (POD logic is copyable by Printify/Printful in a quarter). Days-to-revenue dinged because the kill criteria correctly anticipate a 30-day window, and aspirational-yes risk on Reddit is real. Solid low-risk probe but ceiling concerns are legitimate.
Reply "approve #2" on Telegram to ship this bet.
Bet #3

Private malpractice coach for junior lawyers

judge 76/100edge 1.0/10info product

Picture a third-year associate at a regional firm who just missed a filing deadline. She can't ask her risk partner what happens to her career, can't tell if she's even covered under the firm's umbrella, and can't decode whether 'claims-made' means she's exposed after she laterals. Every existing resource — carrier PDFs, bar association memos, $300 CLEs — is written for firm administrators, not the 26-year-old whose career is actually on the line.

There are ~100k junior lawyers in the US. At a 1% capture and $25 ARPU, that's $300k ARR; at 2% and $30, it's $720k. Nobody is selling them an interactive scenario tool today — the wedge is real, the privacy angle is sharp, and willingness-to-pay is the open question.

I won't oversell the moat: the IP is a decision tree, and a bar association could clone it in 60 days. We also have no operator edge here — this isn't manufacturing or wine. So we treat it as pure demand validation. Why now: r/Lawyertalk and legal Twitter have made associates comfortable seeking advice outside their firm.

The test: a free 2-minute fear-assessment quiz, distributed through r/Lawyertalk and law school alumni lists, upselling to a $29 PDF and a $19/mo subscription. 14 days, ~$200 in ad spend, kill at <75 starts or <5% PDF conversion, hard kill at 200 starts with zero paid by day 45.

It's cheap, it's fast, and the pain is real even if the trigger moment is fuzzy. Let's find out if abstract peace of mind opens a wallet.

The detail behind the pitch
Problem
Junior lawyers don't understand what malpractice insurance covers (defense costs vs. judgment/settlement), how claims affect future employment, or what to do after a mistake.
Proposed solution
An interactive guide/calculator that shows malpractice claim outcomes, insurance coverage specifics, reporting obligations, and firm/career impact based on mistake severity and firm size.
Target market
Junior lawyers (1-5 years licensed); ~100k in US; willing to pay $15-40/month for peace of mind and clarity.
First test
Create a free 2-minute quiz that assesses common malpractice fears, then upsell to a detailed PDF guide. Distribute via r/Lawyertalk and law school alumni groups. Measure: quiz completion and PDF downloads in 14 days.
Kill criteria
<75 quiz starts in 14 days OR <5% of quiz completers download the PDF guide AND 0 paying conversions to any upsell by day 30 → kill; alternatively, >200 quiz starts but <3 paid conversions at any price point by day 45 → kill (validates interest, kills willingness-to-pay thesis)
Competitive landscape
Incumbents: Lawyers Mutual Insurance, ALPS (Attorneys Liability Protection Society), Chubb Lawyers Professional Liability, LawPay (adjacent fintech), Westport Insurance, PLI (Practising Law Institute) — CLE courses on malpractice Pricing: No direct interactive-tool competitor found; CLE courses on malpractice risk run $50-$300 one-time; malpractice insurance itself runs $1,200-$3,500/yr for solo/junior coverage; no $15-40/mo SaaS calculator targeting junior associates exists in training data Saturation: low Wedge: Zero interactive, scenario-based tools exist for individual junior lawyers — every existing resource targets firm administrators or risk managers, leaving associates with no personalized, private channel to model 'what happens to me if X goes wrong.' User complaints: Insurance carrier materials are written for firm risk managers, not individual associates — juniors don't know if they're even covered under firm umbrella policies; No plain-language explanation of claims-made vs. occurrence policy distinctions exists for early-career lawyers; Associates don't know their reporting obligations (to firm, to insurer, to bar) after a mistake — timelines are murky; Career impact of a malpractice incident on lateral hiring or partnership track is never discussed openly at firms; Existing guides are static PDFs from bar associations — no scenario modeling or personalization Notes: The space is served only by static bar association PDFs, carrier underwriter guides aimed at firm administrators, and expensive one-off CLE courses — none are interactive, none are priced for individual junior lawyers, and none address career/employment impact alongside coverage mechanics. The privacy angle is strong: associates cannot safely ask their firm's risk partner 'what happens to my career if I make a mistake,' making a private SaaS tool compelling. Primary risk is willingness-to-pay validation — $15-40/mo is plausible for peace of mind but the purchase moment is unclear (onboarding at a firm? After a near-miss?). A bar association partnership or law school distribution channel could sharply reduce CAC.
Skeptic + judge rationale
Death modes: - Junior associates only feel acute pain after a near-miss incident, but the quiz/guide targets them before any incident occurs — 'peace of mind' is too abstract to trigger a $15-40/mo credit card swipe from a 26-year-old making $85k at a regional firm who has never had a malpractice scare; quiz gets completions but conversion to paid stalls at <2% because there's no urgent, time-bound buying trigger - Reddit r/Lawyertalk and alumni list distribution hits a ceiling of ~300-500 quiz starts total (small, skeptical audiences who distrust unsolicited tools), and the 14-day test produces 40 quiz completions and 4 PDF downloads — not because the product is wrong but because the distribution channel has no reach into the actual 100k TAM, killing the signal before it can be interpreted - A bar association (e.g., ABA YLD or state bar) or a free-tier insurance carrier (ALPS, Lawyers Mutual) sees the quiz go modestly viral, clones the concept as a free member benefit within 60 days, and eliminates willingness-to-pay entirely — the moat is zero because the core IP is a scenario decision tree, not proprietary data, and incumbents have the distribution and trust that the founder lacks # Judge rationale (score=76.0) Wins on low capital (pure software quiz + PDF), low ongoing human intervention (self-serve SaaS), recurring SaaS pricing, and a real ~100k buyer pool with a clear privacy wedge. Loses heavily on defensibility — the scenario decision tree is trivially cloneable by bar associations or carriers who own the distribution. Days-to-paid is the soft spot: willingness-to-pay at $15-40/mo for abstract peace-of-mind without an acute trigger moment is the central risk, and conversion likely stretches past 60 days. ARPU is mid-tier; this is a volume play that depends on a distribution unlock (alumni, bar, CLE partnership) that hasn't been secured.
Reply "approve #3" on Telegram to ship this bet.
Bet #4

Domain-aware QA for AI translations

judge 70/100edge 7.0/10infra tooling

HR SaaS teams keep shipping 'Grados de Pago' when payroll engineers in Madrid actually say 'Escalas salariales.' Cosine similarity says 0.92 — passes. Native speakers say 'this is broken' — rework. Every Series A SaaS localizing into 5+ languages is paying humans to catch what AI confidently mistranslates, because Lokalise's QA stops at tag mismatches and length warnings.

Narrow but real: ~1,200 seed-to-Series B SaaS teams localizing 5+ languages, mid-market ARPU around $300/mo for a CI-plugged scoring API. 5% capture × $3.6k ACV = ~$200k ARR floor, with a credible path to $1M if we ride into fintech and legal verticals. Not a unicorn — a sharp wedge.

The wedge is honest-but-thin: domain glossary + multi-engine back-translation as a scoring API that sits in GitHub Actions, no TMS migration. Incumbents have the primitives (Lokalise has glossaries and webhooks) but haven't shipped contextual scoring as a first-class feature. If Lokalise ships it natively in 60 days, we're cooked — that's the real risk and I won't pretend otherwise.

Why now: GPT-4-class back-translation is cheap enough ($0.001/string) to run on every commit, which wasn't true 18 months ago. Why me: I'm bilingual Spanish/English — I can validate domain errors instinctively and sell to LATAM-localizing US SaaS with native credibility.

The bet: 14 days, ~$2k in API spend, 100 HR strings from a design partner. Kill if accuracy lift <20% vs cosine baseline, or <3 CI integrations and <$500 MRR by day 45.

Small, reversible, and I know exactly when to walk. Let's run it.

The detail behind the pitch
Problem
AI localization pipelines produce contextually poor translations that pass similarity thresholds but fail human review (e.g., 'Pay Grades' → 'Grados de Pago' instead of 'Escalas salariales'), causing rework and quality issues for HR software companies.
Proposed solution
Provide a contextual scoring system that combines cosine similarity with domain-aware term mapping and multi-engine back-translation validation to identify bad translations before human review.
Target market
Seed-to-Series B SaaS companies localizing software (HR, fintech, B2B tools) across 5+ languages; estimated 500-2000 teams currently doing this manually.
First test
Build a simple scoring layer that flags translations with <0.85 contextual confidence despite high cosine similarity; test on 100 HR-domain strings from their existing pipeline and measure reduction in human-flagged errors.
Kill criteria
<3 CI/CD pipeline integrations (GitHub Actions or webhook) activated by paying teams AND <$500 MRR by day 45, OR pilot accuracy improvement <20% vs cosine-only baseline on the 100-string HR test by day 14 → kill
Competitive landscape
Incumbents: Lokalise, Phrase (Memsource), Localize.js, Tolgee, Smartling, Transifex, Google Cloud Translation API, DeepL API Pricing: $15-$60/1M chars (MT APIs); TMS platforms ~$120-$500+/mo for SaaS teams; Lokalise/Phrase enterprise tiers unpublished Saturation: medium Wedge: Incumbents provide generic QA (spell-check, tag mismatches, length warnings) but none offer domain-aware scoring that maps industry-specific terminology (HR, fintech) and uses back-translation confidence to surface contextually wrong-but-structurally-valid translations before human review. User complaints: QA tools flag formatting errors but miss contextual/domain terminology mistakes (e.g., HR, fintech jargon translated literally); Steep learning curves on Lokalise and Phrase for non-localization engineers; Existing similarity/QA checks produce false positives — teams still rely on manual human review to catch domain errors; Pricing scales poorly: Lokalise/Phrase integration tiers become expensive quickly, pushing seed-stage SaaS teams to manual workflows; AI suggestions lack app-level context — tools don't know whether a string is HR payroll UI vs. a generic label Notes: The TMS market (Lokalise, Phrase, Smartling, Transifex) is crowded at the workflow layer, but quality scoring sits mostly in add-on LSP services or expensive enterprise plugins, not as a standalone API/scoring layer targeting dev teams. The specific gap — domain-aware contextual scoring for vertical SaaS strings — is underserved; most incumbent QA stops at surface-level checks. The real risk is Lokalise or Phrase shipping a 'glossary enforcement + back-translation' feature as a native capability (Lokalise already has glossary/QA hooks), making this a feature rather than a product. The strongest go-to-market is a scoring API that plugs into existing CI/CD pipelines (GitHub Actions, Lokalise webhooks) so teams don't have to migrate their TMS.
Skeptic + judge rationale
Death modes: - Lokalise ships native 'domain glossary enforcement + back-translation QA' in their existing QA hooks (already in roadmap signals given their glossary/webhook infrastructure) within 60 days, making this a free feature toggle for 50k+ existing Lokalise customers — founder's wedge evaporates before a single paid contract closes - The 100-string HR pilot reduces human-flagged errors by 18% (below the 20% threshold), but even if it hits 25%, the actual buyer (localization engineer or L10n manager) has zero budget authority — procurement requires VP Engineering sign-off, a 3-6 month security/vendor review, and SOC2 compliance that the founder cannot produce, making the sales cycle 2x longer than runway - Free-tier API access (or a generous trial on 1M chars) is consumed by seed-stage teams for their entire localization backlog in one sprint, they churn before hitting a paywall, and MRR stays at $0 because at $15-60/1M chars the unit economics only work at scale that seed-stage companies (the exact target) structurally cannot provide — the 500-2000 target teams are too small to generate meaningful revenue individually and too fragmented to close enterprise deals # Judge rationale (score=70.0) Strong on operational shape: pure software API plugging into CI/CD with subscription pricing, low capital to validate, and Vercel-deployable. Loses heavily on defensibility — Lokalise/Phrase can ship native glossary+back-translation QA and erase the wedge. ARPU is mid (seed-stage SaaS teams won't pay $10k+/yr) and market is narrow (500-2000 real buyers). Human intervention risks creeping up via design-partner pilots and domain-tuning per vertical, though the API shape keeps it from being service-heavy.
Source: hn:ask_hn
Reply "approve #4" on Telegram to ship this bet.
Bet #5

Pre-attorney filter for ChatGPT redlines

judge 71/100edge 1.5/10b2b saas

Family law attorneys are getting ground down by a new tax: clients running draft separation agreements through ChatGPT and showing up with 40 lines of cosmetic redlines that change zero legal effect. The attorney either eats the review time or bills it and looks greedy. Either way, the client thinks they 'caught something' the lawyer missed. It's a trust-eroding, time-eroding mess and it showed up in the last 18 months.

Market is bounded but real: ~30k solo and small-firm family law attorneys in the US, target ARPU $50/mo, 3% capture = ~$540k ARR. Not a unicorn. A respectable, capital-light SaaS line.

The wedge is narrow on purpose. Harvey, CoCounsel, Spellbook all aim at BigLaw transactional work and the attorney seat. Nobody ships a client-facing intake layer for MSAs, parenting plans, and QDROs that triages edits by legal effect before the attorney opens the file. That's the gap.

Honest about the risks: operator-edge fit here is near zero — Lisandro has no legal network, no Spanish-market angle, nothing unfair. And the skeptic's right that ChatGPT redlines may only cost 15 min/matter, not an hour, which collapses willingness-to-pay.

So the test is cheap and brutal: a PDF checklist plus email script, zero code, sent to family law attorneys cold. 14 days. Kill if <5 adopt, <2 report fewer frivolous redlines, or zero will commit $30+/mo when asked directly. Total capital: a weekend and outreach time.

Let me spend two weeks proving the pain is real before we build anything.

The detail behind the pitch
Problem
Family law attorneys are frustrated when clients use ChatGPT to redline agreements, wasting attorney billable time and client money on cosmetic AI rewrites that don't change legal effect.
Proposed solution
A pre-agreement onboarding tool that explains legal vs. cosmetic edits, auto-flags when client-proposed changes are substantive, and gives clients a simple yes/no on whether to proceed with attorney review.
Target market
Solo and small-firm family law attorneys (10-50k in US); $100-200/hr billers; willing to pay $30-80/month to reduce wasted hours on low-value edits.
First test
Create a simple checklist PDF + email script that family law attorneys send to clients before they receive draft agreements. Track: how many attorneys adopt, and do they report fewer frivolous redlines in 14 days?
Kill criteria
<5 attorneys adopt the checklist workflow within 14 days OR among adopters, <2 voluntarily report fewer substantive redlines in writing by day 14 OR 0 attorneys express willingness to pay $30+ when directly asked by day 30 → kill
Competitive landscape
Incumbents: Clio Duo, Harvey AI, CoCounsel (Thomson Reuters), Spellbook (Rally), Ironclad, LawGeex, Paxton AI Pricing: $49-$150/seat/mo (Spellbook ~$99+; Clio bundles vary; Harvey enterprise only; LawGeex enterprise) Saturation: low Wedge: The only client-facing intake layer that classifies edits by legal effect rather than language change, purpose-built for family law document types, priced for solo practitioners. User complaints: Existing tools are built for corporate/transactional law, not family law — no MSA, parenting plan, or QDRO context; Enterprise pricing (Harvey, LawGeex) is completely inaccessible to solo and small-firm family law attorneys; Tools redline for style and language, not substantive legal effect — same core problem the solution tries to solve; No client-facing layer: all incumbents assume the attorney is the user, leaving the client-side chaos unaddressed; Attorneys still must review every AI suggestion line-by-line; no triage or prioritization by legal significance Notes: The redline-review space is crowded at the enterprise and transactional level, but family law is genuinely underserved — incumbents (Harvey, CoCounsel, LawGeex) target BigLaw and in-house teams with contract volumes that don't resemble family law practice. The client-side angle is the real whitespace: no incumbent positions a tool as a pre-attorney-review filter that educates and triages the client, which is the exact pain point described. The $30-80/month price target is realistic and under-served; Spellbook is the closest competitor but is transactional-contract-focused and attorney-facing only. Key risk is that the market is small (10-50k solo family law attorneys) and conversion/willingness-to-pay may be soft unless the tool demonstrably saves a billable hour per client per matter.
Skeptic + judge rationale
Death modes: - Attorneys don't send the checklist PDF because client onboarding is chaotic and adding a pre-agreement step feels like more work than the ChatGPT redlines cost them — adoption stalls at <5 attorneys in 14 days because the cure requires more behavioral change than the disease causes pain - The 'substantive vs. cosmetic edit' classification fails on family law edge cases (e.g., a custody pickup-time change that looks cosmetic but has enforcement implications), causing one attorney to send a misleading yes/no to a client who then skips attorney review on a genuinely material change — attorney liability fear kills referrals and word-of-mouth cold - Solo family law attorneys earning $100-200/hr calculate that a $49-80/month SaaS subscription requires saving only 0.5 hrs/month to break even, but when surveyed post-trial they report ChatGPT redline reviews average 15 minutes per matter, not the assumed hour — perceived ROI collapses and renewal rate at day 60 is near zero # Judge rationale (score=71.0) Wins on capital (PDF + email script costs nothing) and recurring SaaS model with a clear wedge in an underserved family-law niche. Loses on ARPU ($30-80/mo is mid-tier), market size (10-50k solo attorneys is bounded), and defensibility (classification logic is replicable once Spellbook or Clio notices). Human intervention is the real risk: attorney onboarding, edge-case liability handholding, and family-law-specific classification tuning likely pull Lisandro into 5+ hr/week support during the first 90 days. Skeptic's adoption-friction and ROI-collapse modes are credible — the disease may genuinely be smaller than the cure.
Reply "approve #5" on Telegram to ship this bet.

★ Killian's Wildcard

Off-Brief, Off-Hand

Tonight's instinct bet — synthesized from training, not pulled from sources. Same calibration, different lane.
Killian held the column tonight. None of the instinct candidates cleared the judge floor — better silence than a weak bet. Tomorrow, again.