Evaluation suites for the MimicScribe meeting transcription pipeline.
| Benchmark | What it measures | Results |
|---|---|---|
| Diarization | Speaker attribution accuracy on public corpora (AMI, Earnings-21, VoxConverse, SCOTUS) | Results |
| Meeting Assistant | Real-time briefing quality — talking points, action items, question detection, interpersonal awareness, long-meeting compaction | Results |
| Context Retrieval | Reference document retrieval + end-to-end hallucination safety across 15 document types — CRM, HTML, PDF, contracts, competitive intel, strategic plans | Results |
| Meeting Search | Semantic search across months of meetings — buried sub-topics, disambiguation, ASR noise tolerance, conversational vs formal phrasing | Results |
| Transform | Selection-based text transformation quality — number/date/URL preservation, anti-injection from Reference Documents, voice matching, collision resilience | Results |