Why M4A matters for mobile-first workflows
The vast majority of voice content in 2026 is captured on phones, and the vast majority of phones default to M4A or its near-cousin AAC. iPhone Voice Memos = M4A. Most podcast apps that record interviews on iPhone = M4A. iOS Music app rips = M4A. Demanding "convert to MP3 first" adds a step that loses quality (re-encoding lossy audio is always slightly worse than the original) and breaks momentum. Direct M4A support means the file goes from your iPhone to Markdown without a round-trip through Audacity or an online converter.
Common iPhone-recording use cases
Field interviews: journalist hits Record on the iPhone, conducts a 45-minute interview, AirDrops the M4A to laptop, drops it into MDisBetter, has a structured transcript with speaker labels in 2 minutes. Meeting capture without a bot: place the iPhone on the conference table, hit Record, end the meeting, transcribe the M4A — no calendar invite to a meeting bot, no recording disclosure beyond what your jurisdiction requires you to handle yourself. Voice notes for self: rambling 5-minute idea dumps while walking, transcribed into searchable Markdown notes for later review.
Speaker labels and structure
iPhone Voice Memos are typically captured in noisy environments with one or two voices in close proximity. The transcription engine handles two-speaker conversations cleanly with positional labels (**Speaker 1:**, **Speaker 2:**); for very noisy environments or 3+ overlapping voices, accuracy degrades and you may need to spot-check the labels. Timestamps mark each speaker turn so you can jump back to the audio if a passage needs review.