Step 01
Connect read-only support sources
Point Meihaku at the docs, tickets, macros, and notes your team already relies on.


AI-assisted readiness map
AI maps every customer intent against connected support sources, then shows cited gaps, conflicts, partial coverage, and answers ready for approval.

Conflict
Two sources disagree.
Gap
No approved guidance found.
Partial
One required condition is missing.
Ready
Evidence is cited and reviewed.
Product overview
Meihaku uses AI to compare support sources against customer intents. Reviewers see citations, fix gaps or conflicts, then approve answers for export.
01
Map
Customer intents
02
Check
Coverage and conflicts
03
Resolve
Gaps and partial answers
04
Export
Approved answers
Step 01
Point Meihaku at the docs, tickets, macros, and notes your team already relies on.

Step 02
See which customer questions are cleared for AI, blocked, conflicted, or must route to a human before your AI support agent handles them.

Step 03
When two sources disagree, reviewers see both claims side by side and choose the canonical answer.

Step 04
Only source-backed answers become clearable for the downstream agent your team is about to launch.

Choose the next move
Start from the decision in front of you: build the readiness map, find source blockers, or prepare a vendor-specific launch.

Launch in progress
Connect one source bundle and see which customer intents are ready, partial, missing, or conflicting.
Start readiness audit
Source quality
Check source coverage, conflict risk, escalation readiness, governance, and wrong-answer measurement.
Score readiness
Vendor rollout
Pick the platform your team is launching, then use the vendor-specific checklist, source path, and audit workflow.
Choose platformWhat the audit does
Most teams discover the real work after the chatbot project starts: outdated docs, conflicting policies, missing answers, and tribal knowledge trapped in tickets.
Start with the questions your AI support agent is expected to handle. Meihaku groups them into topics and intent checks.
Meihaku scans read-only support source bundles, then marks each intent as ready, partial, gap, or conflict with citations.
Your team adds missing answers, picks canonical sources, approves cited guidance, and exports only approved content.
Sample audit output
Each customer intent gets a status, source evidence, blocker, and approval decision your team can act on before export.
Score your readinessReady
Refund policy and Zendesk macro agree on 30 days.
Approve answer with citation.
Conflict
Help center says 10 business days; macro says 14.
Pick the canonical source before export.
Gap
No approved source explains the verification path.
Add the missing answer or keep unsupported.
Source and data boundary
Buyers need to know what AI does, what stays read-only, and where human approval gates the downstream bot.
Review securitySource access
Review the policies, macros, SOPs, FAQ docs, ticket examples, and internal notes your support team already uses.
AI role
AI compares source content against customer intents and shows the citations behind every readiness status.
Governance
Answers only move downstream after your team resolves blockers and approves the cited guidance.
Why readiness comes first
If the source knowledge is missing, stale, or contradictory, your AI agent can still answer with confidence. Meihaku turns source risk into explicit decisions: cleared for AI, human handoff required, blocked topic, or source fix needed.
Air Canada
Tribunal ordered the airline to honor a bereavement-fare policy its chatbot invented. C$812 + ruling that the chatbot is part of the website.
Moffatt v. Air Canada, BCCRT 2024 →Cursor
Support bot “Sam” invented a one-device-per-subscription policy that did not exist. Subscription cancellations followed.
Fortune, April 2025 →DPD
Chatbot swore at a customer and wrote a poem calling DPD “the worst delivery firm in the world.” 1.3M views on X.
TIME, January 2024 →“Air Canada is responsible for all of the information on its website, regardless of whether said information comes from a ‘static’ webpage or a chatbot.”
Deeper launch artifacts
Use the platform pages, templates, tools, and guides when you need to investigate a specific vendor, policy area, or launch blocker.

Platform paths
Choose the AI support platform or source system your team is about to launch.
View integrations
Interactive tools
Score launch risk and generate the risk register before committing to automation scope.
Open tools
Operator artifacts
Download the launch checklist, batch-test CSV, macro audit, or risk register.
Use templates
Research library
Read the deeper guides on readiness, testing, knowledge audits, platforms, and governance.
Read guidesFrequently asked
Hallucinations usually happen when the model can’t find a clear answer in the source material it retrieves from. Meihaku audits your support docs, tickets, macros, and notes for the gaps and conflicts that force the model to guess, then gates which intents your agent is cleared to answer. Many teams pair Meihaku as the pre-flight check with runtime monitoring as the in-flight safety net.
Three common patterns: (1) gaps — the customer asked something not covered in any source, so the model fills in plausible-sounding nonsense; (2) conflicts — one support source says one thing and another says something else, and the model picks the wrong one; (3) stale or out-of-date content. Meihaku surfaces all three before launch so you can fix them rather than ship them.
A two-layer approach. Pre-flight audits (Meihaku) catch the gaps and conflicts in your sources before the AI agent ever has to retrieve from them. Runtime monitoring tools catch the residual hallucinations that slip through. The fastest single improvement is usually source cleanup — no amount of runtime checking can save you from a knowledge base the model can’t retrieve clearly from.
Yes — and it’s the cleanest case to start with. Meihaku audits your support knowledge before you commit to a vendor. You’ll know in advance which intents are well-supported, which need source cleanup, and which can’t be answered safely by any agent. Most teams find the audit changes which agent they pick, because some handle gap-prone intents better than others.
Meihaku is the readiness layer that runs before and alongside AI support agents, not a replacement. Tools like Intercom Fin, Decagon, Sierra, and Maven generate answers at runtime. Meihaku audits whether your knowledge base actually supports those answers in the first place, surfacing conflicts, gaps, and ungrounded intents before they reach a customer. Most teams use Meihaku to pre-flight an AI agent rollout, then keep it running as a governance layer.
Runtime monitoring tools score AI responses as they’re produced and can block bad outputs in real time. Meihaku works one layer earlier: it audits the source knowledge your AI agent retrieves from, so the bad answer never gets generated in the first place. The two are complementary — Meihaku as the pre-flight check, runtime tools as the in-flight safety net.
We solve the same launch-risk problem from a different layer. Agent QA, simulation, regression, and runtime monitoring tools test how the runtime agent behaves once it is talking to customers. Meihaku prepares the support knowledge before that test: it checks whether the help center, tickets, macros, SOPs, policies, and reviewer decisions support each customer intent before any agent is allowed to answer it. Many teams need both: Meihaku to define the approved answer boundary, then agent QA tools to test how the runtime agent behaves inside that boundary.
Partly, but the audit unit is not a document. It is a customer intent. Meihaku reads support docs, tickets, macros, SOPs, and policies, then checks whether each customer question has one current, cited, customer-safe answer. That makes it different from general document audit, help-center CMS, or public-page AI-readiness scanners: the output is an approved answer boundary for AI support, not just cleaner documentation.
Meihaku is designed to review the support sources your team already relies on: policies, SOPs, macros, FAQ docs, ticket examples, and internal notes. Source access stays read-only, and Meihaku never writes back to those systems — what you see in the audit is exactly what your team approves.
Yes. AI does the first pass: it maps customer intents to your support sources and flags gaps, conflicts, partial coverage, and ready answers. The result is not a black-box approval. Each status includes evidence for your team to review, and only human-approved answers are exported downstream.
No. Meihaku is complementary. Your AI agent (Fin, Decagon, Sierra, custom) still answers customers. Meihaku decides which intents are cleared for that agent to handle, with cited evidence, and exports the approved answer set. Unsupported or conflicting intents stay out of automation until your team has source evidence or a handoff path.
Yes — that’s the core workflow. Meihaku drafts cited answers from your sources, your team reviews each one with the source line attached, and only cleared intents move toward automation. Nothing reaches production until you greenlight it.
When the same intent (for example, the refund window for international orders) resolves to different answers across support sources, Meihaku flags it as a conflict and shows both source lines side-by-side. Your team picks the canonical answer, which becomes the approved version exported to your AI agent.
That’s the most common starting point — and exactly what Meihaku is designed for. You don’t need to rewrite anything before connecting. Meihaku reads what you have, drafts cited answers from the strongest evidence, flags gaps and conflicts, and lets your team triage them in priority order. The first audit usually surfaces the small set of intents blocking most of the customer-risk conversations.
Connecting a read-only source takes minutes when credentials are ready. Meihaku can generate the first intent map quickly, then source scanning time depends on content volume. Most teams should start with one high-risk folder or ticket bundle before expanding the review.
No. Customer data is processed in your isolated workspace. We don’t train foundation models on your content, and we don’t share data across customers. Source content is processed transiently and not retained beyond what’s needed for citations.
Let AI map intents to evidence, then let your team resolve blockers, approve guidance, and export only what the bot is allowed to use.