Managing partners at most US law firms still believe Google is the primary research surface their attorneys use. The data inside Bing AI Performance suggests otherwise. aivortex.io's last 30 days: 2,100+ Microsoft Copilot citations, top grounding query "Harvey AI legal." Microsoft 365 Copilot reaches 90%+ of US law firms via the M365 install base. Lawyers prompt Copilot inside Word while drafting, inside Outlook while emailing, inside Teams while in meetings — without leaving the document they're working in. Google search requires opening a new tab; Copilot doesn't. That structural difference reshapes where research actually happens. This analysis pulls apart the channel split between Google and Copilot for legal-industry research, with first-party data and the strategic implications most firms haven't priced in.
The structural difference — context-switch cost
Google search has been the default research surface for two decades. Its strength has always been the SERP — a list of ranked sources, the user picks, the user clicks. Its operational cost is the context switch: stop drafting, open a tab, type a query, read results, click, read source, return to the document, integrate the finding.
Copilot collapses that loop. Inside Word, the attorney highlights a clause and prompts "summarize the case law on this exact clause type." Copilot returns a composed answer with citations, inline. No context switch. The user reads the answer in the same document they're editing.
The consequence at scale:
- For tasks where the answer is the value ("what does this rule require?"), Copilot wins because it eliminates the context switch. - For tasks where the source vetting is the value ("what's the most authoritative analysis on this?"), Google still wins because the user reads the source directly. - For tasks in the middle — vendor research, policy summaries, comparison shopping — the surfaces compete. Copilot is winning faster than most firms have noticed.
The Bing AI Performance data on aivortex.io reflects this. Top grounding queries are vendor research and policy questions — exactly the tasks that compress well into a composed answer. Lawyers researching Harvey AI inside Word don't open a new tab. They prompt Copilot. The answer cites Vortex among other sources. The research happens without ever touching Google.
The query types that split between channels
From cross-referenced data — Bing AI Performance, Google Search Console, and direct prompt logs — five query categories show distinct channel behavior:
- Vendor research with niche modifier ("Harvey AI legal," "Spellbook contract review"). Increasingly Copilot. The answer-shaped query maps to Copilot's grounding model better than to a SERP. Vortex's top citation share lives here. - Case-law queries ("Heppner ruling AI privilege," "recent AI hallucination sanctions"). Mixed. Copilot for the summary, Google for the source vetting. Lawyers often run both surfaces in sequence. - Procedural and rule queries ("federal court AI disclosure rules," "state bar ethics CLE requirements"). Increasingly Copilot. The user wants the answer, not a list of sources. Vortex's federal court AI disclosure directory gets cited heavily here. - Brand-name queries (the firm's own name, partner names). Still mostly Google. Brand search is a Google strength because users want the official site, not a composed answer about the firm. - Long-form analysis and journalism ("Above the Law coverage of legal AI," "recent ABA Journal AI articles"). Still mostly Google. Reading long-form content is a Google strength because the value is in the prose itself, not in a summary.
The operational consequence: a firm whose visibility strategy is Google-only is invisible in two of the highest-leverage research categories — vendor research and procedural/rule queries — that managing partners care most about for AI deployment decisions.
First-party data: aivortex.io's split between Google and Copilot
Comparing Vortex's last 30 days across surfaces:
- Bing AI Performance: 2,100+ Copilot citations. Top grounding query "Harvey AI legal." Click-through under 100. Citations dramatically exceed clicks because users read AI answers without clicking to source. - Google Search Console: roughly 50-100 clicks across the same window from organic Google search. Impressions in the 20,000-range, low CTR (~0.4%) — typical pattern for English-language US legal SERPs that Google AI Overviews increasingly eats. - Vercel direct visitors: 486 visitors per week. Mix of organic search, social, and direct.
The ratio that matters: Copilot citations to organic Google clicks runs roughly 20-40x in Vortex's case. The two channels deliver fundamentally different things — citations are AEO presence, clicks are direct traffic — but the AEO presence compounds into authority signals that affect future citations and future Google rankings.
The second-order read: a firm reading only Google Search Console concludes its visibility is modest. The same firm reading Bing AI Performance concludes its visibility is substantial in the AI grounding layer. Both reads are correct for their respective surfaces. Strategic decisions need both data points.
The third-order read: Google AI Overviews and Copilot are converging. Both surfaces compose answers grounded in vetted sources. Content that earns Copilot citations now is structurally identical to content that will earn Google AI Overview citations as that surface matures. The publishing program optimizes for both simultaneously. The Microsoft Copilot citations how-to-rank guide covers the production specifics.
The managing-partner research workflow inside Copilot
What does a managing partner's research workflow look like in 2026 when they're evaluating an AI tool?
Observed pattern across multiple firm conversations:
- Open Word for the existing draft (an AI committee memo, a vendor evaluation document, a policy update) - Highlight the topic in the document, prompt Copilot "give me the current state of [tool] for AmLaw firms" - Read Copilot's composed answer with cited sources - For high-trust sources cited — typically Above the Law, Law360, the firm's preferred industry analyst, or a verified vertical site — click through to read the original - For low-trust or unfamiliar sources — generic SEO content, vendor marketing — skip the citation - Integrate the high-trust findings into the document, prompt Copilot to summarize key tradeoffs - Write the next paragraph based on the synthesized read
The workflow happens inside Word. Total elapsed time: 15-30 minutes for a topic the partner had no prior context on. Compared to the pre-Copilot workflow (open Westlaw or Lexis, search, read, paste citations into a Word document, switch back, draft) — substantially compressed.
The operational consequence for content publishers: being a Copilot citation source is high-leverage because the partner sees the citation but rarely clicks through. The brand impression is the citation itself, alongside the answer. A firm cited 70 times per day inside Copilot has earned 70 brand impressions per day with managing partners actively researching adjacent topics — without paying for any of them. That's the channel arbitrage. The why most firms are invisible inside Copilot analysis covers the strategic implications of missing this channel.
Cross-channel strategy — where to put publishing dollars
The optimization isn't Google vs Copilot. It's how to allocate publishing investment across both, given they reward overlapping but distinct content patterns.
What optimizes for both: - FAQ-first structure with self-contained answers - Clean schema (Article, FAQPage, BreadcrumbList, Person) - Named authorship with verifiable sameAs links - Current citations against authoritative sources - Internal linking to topical clusters
What optimizes more for Google: - Long-form deep analysis (1,500-3,500 word anchor pages) - Original data and case studies - Editorial voice, brand differentiation - Backlinks from industry publications - Date-stuffed timely content for date-sensitive queries
What optimizes more for Copilot: - Tight per-question depth (1,200-1,800 word spoke pages) - Cluster anchor + spokes pattern with 5-25 spokes per cluster - IndexNow on every publish for Bing crawl acceleration - Vertical specificity over generalist breadth
Most firm publishing programs in 2026 should run an 80/20 split — 80% of pages in spoke format optimized for Copilot citation, 20% in anchor format optimized for Google authority. The two reinforce each other. Anchor pages collect Google authority that helps spoke pages get crawled and cited. Spoke pages drive Copilot citations that feed back as authority signals to anchors. Compare against the Bing AI Performance dashboard guide for the measurement-side workflow.
Recommendations by firm size
Solo and small firms (2-10 attorneys). Don't try to optimize for both channels with limited content investment. Pick the channel that matches your client's research behavior. If your clients are corporate decision-makers using M365, Copilot is the higher-leverage channel because that's where they prompt for vendor research and outside counsel evaluation. If your clients are consumer or media-driven (PI plaintiffs, criminal defense families), Google still dominates because the discovery happens via SERP.
Mid-size firms (10-50 attorneys). Run the 80/20 spoke-anchor split described above. 50-75 pages per quarter is the cadence that compounds within 12 months. Open both Bing AI Performance and Google Search Console monthly. Different metrics, both required for full visibility.
BigLaw and AmLaw 100. Run programmatic publishing across multiple clusters with both channel optimization built in. Tie content production into KM and marketing analytics. Compare your channel split against this analysis quarterly. The firms doing this well today will own the AEO citation share in 2027-2028 when most competitors finally notice. Pair the data read with the Microsoft Copilot citations how-to-rank guide for the production-side specifics.
The Bottom Line: My take: Managing partners increasingly research inside Copilot, not Google, because the context-switch cost is lower. Vortex's first-party data shows 20-40x more Copilot citations than Google clicks on the same content. Firms whose visibility strategy is Google-only are invisible in the highest-leverage research surface their own attorneys and clients now use daily.
AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.
