
Introduction
If you've spent the last year publishing well-structured, well-sourced, comprehensively researched content and you still aren't getting cited in ChatGPT, Perplexity, or Google's AI Overviews, the problem probably isn't your structure. It is that your content doesn't add anything the rest of the internet doesn't already say.
That's the bet behind a concept called information gain — the new metric several AEO practitioners are using to describe the gap between what your site already covers and what only your business can credibly add. The argument is straightforward: AI systems are increasingly choosing between sources that all cover the topic well. The tiebreaker is whichever source contributes a fact, framing, or data point the others didn't.
A practical framework for finding those gaps was published on LSEO on April 21, 2026 by Kristopher Jones. It is targeted at enterprise content teams, but the underlying audit translates cleanly to a small business that can run it in a weekend. This guide adapts that framework into a six-step checklist a non-technical owner-operator can complete with a CRM export, a notepad, and a few hours of focused time.
The Northeast Indiana angle matters here because most Fort Wayne and Auburn small businesses are sitting on more proprietary data than they realize. The data is in the call logs, the intake forms, the service tickets, and the seasonal calendar. The challenge is recognizing what's worth publishing and what isn't.
Key Takeaways
- Information gain is the gap between what your site already says and what only your business can credibly add
- AI systems pick between competing sources by looking for distinctive contributions, not just comprehensive coverage
- A six-step audit can be run in a weekend by a non-technical small business owner using existing CRM and call data
- There are two types of proprietary data: hard data (analytics, CRM, operational metrics) and soft data (expert heuristics, recurring objections, local context)
- Northeast Indiana small businesses already hold the data — the gap is usually publishing it with honest methodology, not collecting it
- The audit is meant to be repeatable, not a one-time project; rerun it quarterly as the public conversation shifts
What Is Information Gain and Why Does It Matter Now?
The original LSEO definition is precise: information gain measures the gap between what your website already says, what competing sources already cover, and what unique knowledge your business could contribute but has not yet published. It is not about volume of content. It is about contribution per page.
That distinction matters because most SMBs publish what their competitors publish, structured the same way, often using the same source material. When AI systems generate an answer, they look across that pool of similar pages and pick whichever ones add the most distinct signal. Pages that repeat what's already in the corpus get filtered out before the citation step.
A study published in Search Engine Land on April 16, 2026 analyzed 50,553 ChatGPT responses across 16,851 unique queries and found that pages between 500-2,000 words earned citations more reliably than longer ones — and pages over 5,000 words were cited less often than pages under 500. Length isn't the goal. Distinctiveness packed into a tight answer is. We unpacked that data in detail in ChatGPT citations favor ranking and precision over length, and the implication for content strategy is clear: a short page that contributes one verifiable proprietary fact will reliably out-cite a long page that doesn't contribute anything new.
The honest version of this is that information gain is harder for small businesses to fake than coverage was. You can spin up a 2,500-word topic-cluster post in a few hours with the help of an AI writing tool. You cannot fabricate proprietary data without lying about your own business. The good news is that most SMBs don't need to fabricate anything — they already have the data they need. They just haven't published it in a form an AI system can extract.

What Counts as Proprietary Data for a Small Business?
The LSEO framework distinguishes two categories that map directly onto what a Fort Wayne service business or B2B manufacturer already collects.
Hard proprietary data is structured information from your systems: CRM records, analytics, call-tracking logs, scheduling data, payment records, support tickets. Most small businesses hold years of this without ever using it for content. Examples for a typical Northeast Indiana service business:
| System | Data already there | Citation-worthy framing |
|---|---|---|
| CRM | Average days from inquiry to closed job, by service type | Service-specific timing benchmark for your category |
| Call tracking | Top 25 questions asked before booking, by season | Decision guide answering the actual pre-purchase questions |
| Scheduling | Average response time, by day of week | Reliability commitment customers can verify |
| Payment records | Typical project size by service category | Honest budget guide for prospective customers |
| Support tickets | Most common post-job clarifications | “Things we wish we'd told you” article |
Soft proprietary knowledge is everything experienced practitioners know that hasn't been written down. This includes recurring customer objections, technical heuristics that aren't in any textbook, regional patterns specific to your service area, and lessons from work that didn't go as planned. The LSEO piece notes that soft knowledge often produces stronger position signals than hard data because it can't be reverse-engineered from public sources — but it requires more deliberate effort to surface.
Both categories share a property the framework calls “decision usefulness” — the test of whether a piece of content gives someone enough to make a real decision, not just understand a topic. The same logic appears in our why your content doesn't appear in AI Overviews post: AI systems prefer pages that contain specific, actionable information over pages that explain a concept abstractly.
The constraint that catches most operators off guard is the ethical one. The LSEO framework explicitly calls for minimizing, anonymizing, and aggregating customer data; including methodology metadata; and never overgeneralizing findings beyond actual scope. We come back to that constraint in the audit steps below — it is the difference between a citation-worthy contribution and a privacy or compliance problem.
What Are the Six Audit Steps?
The full LSEO framework includes governance, capture, validate, publish, and measure stages that suit an enterprise content team. We've adapted it into six steps a small-business owner can run in a long weekend. Each step has a single deliverable and a defined output.
Step 1: Inventory your top ten existing pages. List your ten highest-traffic or highest-converting pages. For each, write down the topic in one sentence and the single most distinctive claim on the page. If you can't identify a distinctive claim, the page is at the coverage layer only — flag it for the audit.
Step 2: Pull recent customer interactions. Export the last 90 days of call logs, intake forms, chat transcripts, and support tickets. Don't filter yet — you want the raw stream of what real prospects and customers are actually asking. For most Fort Wayne service businesses, this is a few hundred to a few thousand interactions; for a smaller practice, it may be 50-100. Either way, it's enough.
Step 3: Cluster the questions. Group the interactions into 10-20 recurring themes. The goal isn't statistical precision — it's pattern recognition. By the end of this step, you should have a short list of questions or concerns that come up repeatedly, ranked roughly by frequency.
Step 4: Identify the gap. For each cluster, ask: is this question answered well on my site, on a competitor's site, or anywhere on the public internet? The clusters where the answer is “no” are your information gain opportunities. The clusters where the answer is “yes, on competitor sites but not mine” are your coverage gaps — they need to be filled, but they aren't distinctiveness contributions.
Step 5: Pick three contributions to publish this quarter. Not all of them — three. Pick the three with the highest frequency in your interaction data and the lowest existing public coverage. For each, define the format (FAQ, decision guide, benchmark article, methodology page) and the data you'll need to support the claim.
Step 6: Draft, validate, publish. Write the contribution honestly. Include methodology — the source system, date range, sample size, and limitations. Have someone other than the writer verify the data before publishing. Add appropriate schema. Link the new content from a relevant existing page.
That's the audit. The output is a short list of three contributions to publish over the next quarter, each grounded in data your business actually has. We've published a related framework for finding keyword gaps using Search Console in intent gap analysis using Search Console; the information-gain audit complements it by working from your private data instead of from search behavior.

How Do You Make the Data Verifiable Without Hiring a Statistician?
This is where most operators get stuck. The LSEO follow-up piece on verifiable claims (April 24, 2026) lays out what makes a claim machine-extractable: a metric, a subject, a method, a timeframe, a comparison point, and a source. Their example: instead of “fast onboarding,” say “average onboarding completed in 11 days across 214 mid-market accounts in 2024.”
For a small business, the verifiable-claim format is achievable without a statistician. The structure is:
- Metric: what you measured (average response time, project duration, win rate)
- Subject: who it applies to (customers in DeKalb County, residential customers, B2B accounts under $50k)
- Method: how you measured it (CRM timestamp from inquiry to first contact, scheduling system from job start to job end)
- Timeframe: when (last 12 months, 2024-2025, since the last operational change)
- Comparison point: what reference frame helps the reader (industry average, your own prior period, a benchmark from a credible source)
- Source: where the data lives (your CRM, your scheduling system, an aggregated export)
The honest tradeoff is that this format requires you to be specific. “We respond fast” is easier to write than something like “average response time was [X] minutes across [N] inbound calls in [time period], measured from call receipt to first callback in our call-tracking system.” The second one is harder to write — but a sentence in that shape, populated with your real numbers, is the one AI systems will actually cite. We covered the broader strategy of building this kind of evidence into your content in our answer engine optimization guide; the information-gain audit is the practical workflow that produces the underlying numbers.
For the schema layer, Schema.org Dataset is the most semantically appropriate type when you publish a discrete data block, while Schema.org Article with a clear dateCreated and named author is the right wrapper when the contribution is more narrative. Both should reference your Organization schema by @id so the entity attribution is unambiguous.
What If You Don't Have Hard Data — Just Expert Knowledge?
Most owner-operated small businesses are richer in soft proprietary knowledge than in clean structured data. The LSEO framework on expert interview workflows (April 23, 2026) is built for this case, and the structure adapts well to a 30-minute interview an owner can do with their own most experienced technician, attorney, dentist, or installer.
The mechanic is straightforward. Pick one experienced person in the business. Prepare 5-7 questions in advance, layered from definitions to misconceptions to decision criteria. Record the conversation (with consent), transcribe it, and convert the raw transcript into multiple modular assets:
- A verified-quote bank for use in service pages and FAQ answers
- A misconception list for an “X myths” article or FAQ section
- A decision framework that can power a comparison or “how to choose” page
- Real scenario examples that illustrate complex advice
- Risk-disclosure language that improves trust without overstating capability
The LSEO piece's principle — “original expert input is not just a trust signal; it is a production system” — is the right framing. One thirty-minute interview can fuel five or six pages over a quarter, each adding distinct information that no competitor can replicate because the underlying expert isn't on their staff.
For a Fort Wayne home-services business, this might be a half-hour with the senior installer about which 1970s-era furnace models are still common in older Fort Wayne neighborhoods and what that implies for replacement decisions. For a regional law firm, it's a half-hour with the senior attorney about which Allen County court procedures differ from neighboring counties. The contribution is local, specific, and structurally impossible to fabricate from public sources.

What Does This Look Like for a Northeast Indiana Small Business?
A concrete worked example, with names removed: imagine a Fort Wayne home-services business — let's say HVAC and plumbing, two service categories, a 12-person team, six trucks, serving Allen, DeKalb, and Whitley counties.
The audit, run on a real Saturday morning, surfaces something like this. The CRM shows that “average time-from-call-to-appointment” varies meaningfully by service category — emergency calls are dispatched the same day, routine maintenance averages around five business days, full-system replacements average two to three weeks. The call log shows that the most-asked pre-booking question, by frequency, is some variant of “do you service my township.” The intake form shows a recurring concern about how much an HVAC replacement actually costs, with most callers asking before any technician visit.
That's three information-gain opportunities pulled from data the business already had. The publishing step turns each into a piece of content:
- A “what to expect” page that publishes the actual response-time ranges by service category, with methodology (calculated from CRM dispatch data over the last 12 months, sample size, limitations like “weather emergencies excluded”).
- A clear, accurate service-area page or service area FAQ that lists every township and zip code served, with a clear note about which are within standard dispatch range and which carry an extended-area surcharge.
- A budget-range guide for HVAC replacement that publishes the actual price band the business charges, with methodology (range based on installations completed in 2025, factors that move price up or down, why the high end exists).
None of those three contributions has hyperbole. None has a fabricated statistic. Each one says something specific that a competitor cannot copy without their own data. And each one is the kind of contribution AI systems can extract and cite when a Fort Wayne neighbor asks ChatGPT “how much does HVAC replacement cost in Allen County.”
The qualitative claim — “hyper-local content beats generic content” — has been in the AEO literature for at least a year. The information-gain audit is what turns that claim into a repeatable workflow. Our 3 AI-driven SEO frameworks small businesses can run post covers three other workflows that pair well with this audit; the information-gain audit slots in as the input that gives the other frameworks something distinctive to optimize.
If you'd like a structured pass at this for your own business — including the CRM extraction, the cluster analysis, the methodology pages, and the schema work — our content marketing services include an information-gain audit and a quarterly publishing plan built from your own data.

How Do You Test Whether Your Contributions Are Actually Working?
The fourth LSEO piece in the series, on A/B testing for AEO (April 21, 2026), makes a useful argument: testing for citation requires more discipline than testing for clicks because the measurement layer is harder to read. Each variation has to be independently publishable and factually sound — you can't run a “fake number” variation against a “real number” variation, because the fake one would be dishonest content even if it tested better.
For a small business, full A/B testing on AEO is usually overkill. A simpler signal works: track citation share for your information-gain contributions across ChatGPT, Perplexity, and Google AI Mode by running 5-10 sample queries monthly for the questions the contribution is meant to answer. After 60-90 days, you can see whether the contribution is being cited and by which surfaces.
If a contribution earns no citations after 90 days, the LSEO framework points to four common causes: the page's evidence quality is weak, the methodology isn't visible, the page isn't structured for machine extraction, or the contribution is genuinely too niche to surface in any natural-language query. The fix depends on which one applies — but knowing which one applies requires looking at the page honestly rather than blaming “the algorithm.”
The harder honest tradeoff: information-gain work is slower than coverage work. A coverage post can move metrics within a few weeks. A genuine information-gain contribution often takes one to three quarters to compound into measurable citation share. The compounding is more durable, but the patience requirement is real. Operators who want fast wins should focus on coverage and structural fixes first; operators who want durable distinctiveness should run the audit.
Where to Start If You Have a Weekend
If you have one Saturday, run steps 1-4 of the audit. List your top ten pages, pull the last 90 days of customer interactions, cluster the recurring questions, and identify the three biggest gaps. The output is a list, not yet content. That list is enough to know whether the audit is worth a second weekend.
If you have a long weekend, add steps 5-6: pick three contributions, draft them with explicit methodology, and publish them with appropriate schema. You don't need to publish all three at once — one well-built contribution is more valuable than three rushed ones.
If you'd like a second pair of eyes on the audit before committing to publishing, contact us. We'll do a 30-minute pass on your top three pages and your call-log themes and tell you honestly whether the information-gain gap is large enough to justify the publishing work. For some businesses, the answer is “your existing pages are already distinctive — fix the schema and move on.” For others, the audit reveals the missing piece that's been keeping their AEO investment from compounding.
Sources & Further Reading
- LSEO: lseo.com/information-gain-audits-identifying-gaps-in-proprietary-data — Information Gain Audits: Identifying Gaps in Proprietary Data (April 21, 2026)
- LSEO: lseo.com/verifiable-claims-using-quantified-evidence-to-influence-ai-logic — Verifiable Claims: Using Quantified Evidence to Influence AI Logic (April 24, 2026)
- LSEO: lseo.com/expert-interview-workflows-generating-original-data-for-ymyl-aeo — Expert Interview Workflows: Generating Original Data for YMYL AEO (April 23, 2026)
- LSEO: lseo.com/a-b-testing-for-aeo-testing-summary-variations-for-citation — A/B Testing for AEO: Testing Summary Variations for Citation (April 21, 2026)
- Search Engine Land: searchengineland.com/chatgpt-citations-ranking-precision-length-study-474538 — ChatGPT citations reward ranking and precision over length (April 16, 2026)
- Schema.org: schema.org/Dataset — Schema.org Dataset type
- Schema.org: schema.org/Article — Schema.org Article type
- Google Search Central: developers.google.com/search/docs/fundamentals/creating-helpful-content — Google Helpful Content Guidelines
Ready to find your information gain gaps?
Button Block runs information-gain audits for Northeast Indiana small businesses, including CRM extraction, cluster analysis, methodology pages, and schema work. We'll tell you honestly whether the gap is worth closing before you commit to publishing.
Start the Conversation