The Iceberg Problem: Why AI Can’t Build You a Reliable Family Office List

The Iceberg Problem: Why AI Can’t Build You a Reliable Family Office List

There is a question that surfaces in almost every conversation about AI and private markets intelligence: “Couldn’t we just use ChatGPT for this?” It is a reasonable question. The answer, it turns out, depends entirely on what you mean by “this” — and the distinction matters a great deal more than most people realise.

The Iceberg Problem in Family Office Data

Imagine the global family office universe as an iceberg. The visible portion above the waterline — offices that have given interviews, co-sponsored conferences, published investment theses, or appeared in deal announcements — is real, but it represents a fraction of what is actually there. The majority sits below: offices that have always operated quietly, offices that were once visible but have since retreated from public view, and offices whose most important characteristics (investment appetite, ticket size, decision-maker relationships) were never published anywhere at all.

AI systems are excellent tools for mapping the part of the iceberg above the waterline. They are, by design, incapable of seeing the rest. This is not a limitation that better models or more compute will overcome — it is a consequence of what those systems are trained on and how information flows (or doesn’t) in the family office world.

The practical consequences of this distinction become very concrete when you try to use AI to build a comprehensive family office list from scratch.

“AI can tell you about the family offices that wanted to be found. A curated database can tell you about the ones that didn’t.”

We Modelled the Attempt. Here Is the Arithmetic.

Say the goal is a list of 100 verified single family offices in the United States — with investment mandates, approximate AUM, and enough detail to be useful for origination or relationship development. How would you actually do this systematically with an AI API? And what would it cost?

There is no shortcut to a reliable answer. You would need to work through at least three major data streams, plus a structured profiling phase for every candidate you surface.

Step 1 — Work the Wealth Lists

The Forbes 400 is the logical entry point. Individuals with personal wealth above $1 billion almost always establish a family office; the question is what kind. Confirming whether a specific billionaire runs a single family office — as opposed to a multi-family arrangement or a private banking mandate — requires targeted research for each name: searching news and filings for entity names, checking EDGAR for registered adviser records, verifying structure, extracting investment focus, and identifying key people. That runs to around 7 queries per person.

400 names × 7 queries = 2,800 queries. Expected yield: ~150 genuine SFO leads.

Step 2 — Scan Transaction Data

Family offices are among the most active — and least visible — buyers in commercial real estate. Identifying them requires monitoring deal flow and working backwards from buyer identity. MSCI Real Assets records roughly 8,000–10,000 U.S. commercial property transactions above $10 million annually.[1] Each requires at minimum two queries to check for family office involvement.

~9,000 transactions × 2 queries = 18,000 queries. Incremental yield: ~80 additional SFO leads.

A similar logic applies to VC and PE deal flow. PitchBook records approximately 20,000 U.S. private market transactions per year, of which roughly 3,000 show documented or probable family office participation.[2] Verifying investor identity for each runs to three queries on average.

3,000 deals × 3 queries = 9,000 queries. Incremental yield: ~60 additional SFO leads.

Step 3 — Profile Every Candidate

At this point you have approximately 200 unique candidates after deduplication — more than you need, but with significant noise. Each requires structured profiling before it becomes usable: confirmation of single-family status, AUM estimate, sector and geographic focus, portfolio positions, personnel, and contact information. Around 8 queries per candidate.

200 candidates × 8 queries = 1,600 queries.

Research Phase Queries SFO Leads Sources Involved
1 Wealth list scanning 2,800 ~150 Forbes 400, EDGAR, news archives, LinkedIn
2a CRE deal monitoring 18,000 ~80 MSCI, CoStar, county deed records
2b VC / PE deal scanning 9,000 ~60 PitchBook, Crunchbase, SEC Form D
3 Candidate profiling 1,600 Entity-specific web research across all above
Total 31,400 ~100–120 Conservative estimate; +30–50% in practice

31,400 queries. Each one carries a token payload — context window, disambiguation instructions, prior findings, structured output requirements. A realistic per-query average: 1,500 input tokens and 800 output tokens. Across the full run: 47.1 million input tokens and 25.1 million output tokens, for a combined total of roughly 72 million tokens.

The Bill, By Model

Model Input cost
/1M tok
Output cost
/1M tok
Total spend Per confirmed entry
(100 SFOs)
Claude Sonnet 4 · Anthropic $3.00 $15.00 $518 $5.18
GPT-4o · OpenAI $2.50 $10.00 $369 $3.69
Gemini 1.5 Pro · Google $1.25 $5.00 $184 $1.85
Average across models $357 $3.57

API pricing at April 2025 standard rates: Anthropic, OpenAI, Google AI Studio. No batch or caching discounts applied.

Those figures assume 100% accuracy on every query — no hallucinated entities, no conflation of holding companies with family offices, no SFO/MFO misclassification. In practice, LLMs working from public data achieve something closer to 70–80% reliability on this kind of structured entity research. Adjusting for that, effective cost per usable entry climbs to $4.50–$7.40 across the three models.

There is also an environmental dimension worth noting: at a conservative 0.003 kWh per query (per Luccioni et al., 2023[3]), the full research run consumes roughly 94 kWh and generates approximately 36 kg of CO₂ against the U.S. grid average — equivalent to driving 150 miles in a standard petrol vehicle, in exchange for a dataset that requires further manual verification.[4]

A note on price trajectories: AI inference costs have fallen sharply and will likely continue to do so. But even modelling an 80% price reduction across all three providers, total spend for this exercise drops to $37–104 — and cost per reliable entry to $0.50–$1.50. By that point, cost is no longer the central objection. Coverage, accuracy, and the irretrievable absence of non-public data remain.

What the Model Cannot See, and Why That Is Permanent

The cost argument is actually the weaker of the two cases against AI-generated family office lists at scale. The stronger one is about knowledge — specifically, about what kind of knowledge exists in the world and where it lives.

A meaningful share of family office intelligence was never digitised in a form that would end up in a training corpus. Some of it was online once and is no longer: offices that appeared in a 2016 conference programme, were mentioned in a regional business newspaper that has since shuttered, or were listed in a directory that no longer exists. Crawling the internet today returns nothing — but those offices are real, active, and investable.

A further layer of intelligence was never published at all. An office’s actual risk appetite, its minimum cheque size, its preference for club deals over fund commitments, the fact that decisions run through one specific family member rather than the official CIO — none of this appears in any source an AI can access. It emerges from conversations, from co-investments, from introductions that led to follow-up calls that led to relationships. It lives in networks, not databases. And it is precisely the intelligence that makes the difference between a cold list and a warm one.

The Real Comparison

When positioned against curated, research-backed alternatives, the AI-generated approach loses on both dimensions that actually matter: cost per reliable entry and depth of coverage. Specialist providers of family office databases — such as the global single family office lists available from dedicated research platforms — typically price a U.S.-focused list of 500+ verified SFOs at around $800.

Approach Cost Verified entries Cost / entry Coverage of non-public data
AI (best-case, Gemini 1.5 Pro) $184 ~100 $1.85 → $2.50 adj. None — public sources only
AI (average across models) $357 ~100 $3.57 → $4.75 adj. None — public sources only
AI (Claude Sonnet 4) $518 ~100 $5.18 → $6.90 adj. None — public sources only
Curated specialist database ~$800 500+ $1.60 Significant — incl. offline sources, relationship data, historical records

Adjusted figures reflect ~75% AI accuracy rate on entity verification. Curated database figure assumes $800 / 500 entries.

At $1.60 per verified entry versus $2.50–$6.90 for AI-generated approximations — and with a fraction of the coverage — the arithmetic of “just using AI” does not hold up at scale.

So When Does AI Actually Make Sense Here?

This is not an argument that AI has no role in family office research. It clearly does — in the right context.

The useful heuristic is scope. For a mandate that requires five to fifteen highly specific names — a particular geography, a defined sector, a minimum AUM threshold — AI-assisted research can return a reasonable shortlist quickly and cheaply, with acceptable error tolerance. The same applies to point-in-time enrichment: adding a specific data field to an existing list, or cross-checking a name against a particular event or announcement.

Where it breaks down is precisely where the investment in curated data pays off: when you need comprehensive market coverage, when the completeness of the list is material to the outcome, and when the intelligence you need has never been, or is no longer, publicly available.

In Summary

  • Building a list of 100 verified U.S. SFOs via AI API requires approximately 31,400 queries and 72 million tokens
  • Total API spend ranges from $184 (Gemini) to $518 (Claude), before accuracy adjustments
  • Effective cost per reliable entry: $2.50–$6.90 across leading models
  • A curated specialist database delivers 500+ verified entries at ~$1.60 per entry — with coverage that AI cannot match
  • Even dramatic future price reductions do not close the data coverage gap — that is a structural, not a pricing, problem
  • AI remains well-suited to small-scale, targeted lookups; curated data is the rational choice for systematic, large-scale work

Sources

[1] MSCI Real Assets (2024). U.S. Capital Trends: Commercial Real Estate Transaction Volume. Threshold: transactions >$10M, institutional-grade assets.

[2] PitchBook Data, Inc. (2024). Annual U.S. Venture Capital & Private Equity Activity Report. Family office participation estimated at 8–10% of deal volume based on LP disclosure patterns in public filings.

[3] Luccioni, A.S., Viguier, S., & Ligozat, A.L. (2023). Estimating the Carbon Footprint of BLOOM, a 176B Parameter Language Model. JMLR. Per-query energy estimates extrapolated to comparable frontier model families; midpoint of reported 0.002–0.01 kWh range applied.

[4] U.S. Environmental Protection Agency (2024). eGRID 2023 Summary Data. U.S. annual average non-baseload CO₂ output rate: 0.386 kg/kWh.

[5] API pricing as of April 2025. Standard rates, no volume or caching discounts: anthropic.com/pricing, openai.com/api/pricing, ai.google.dev/pricing.

Token and cost figures are modelled estimates based on publicly available pricing and simplified workflow assumptions. Real-world usage will vary based on prompt design, context length, model version, and output complexity.

Comments are closed.