Apple Intelligence transcription is twice as fast as OpenAI's Whisper

Jump to First Reply
Posted:
in iOS

Newly released to developers, Apple Intelligence's transcription tools are fast, accurate, and typically double the speed of OpenAI's longstanding equivalent.

Tablet screen displaying a writing app with a story titled New Story. The text begins with 'Once upon a time in the mystical land of Magenta.'
Apple Intelligence transcription tools being demonstrated by Apple -- image credit: Apple



This could be another example of Apple entering a market late but doing it best. Since OpenAI released its Whisper transcription technology in September 2022, it has become the standard model in dozens of apps -- but now may not stay that way.

For during WWDC 2025, Apple announced that it was opening up its Apple Intelligence tools to third-party developers. While Apple barely gave any details, and certainly did not promote its transcription tools, developers are now discovering what facilities it brings.

MacStories writer John Voorhees and his son spent a reported ten minutes or so, constructing a quick tool to use Apple's transcription technology on the Mac. Specifically, they used elements of Apple's Speech framework called SpeechAnalyzer and SpeechTranscriber, which were released to developers a part of Apple's macOS Tahoe and iOS 26 betas.

This quick tool, a command-line one they call Yap, is available to download from Github. It requires an Apple developer account and macOS Tahoe.

AI transcription tools depend chiefly on the LLM being used, so comparisons are not always simple. However, pitching Apple Intelligence against MacWhisper's Large V3 Turbo model showed a dramatic difference.

Apple's Speech framework tools were consistently just over twice the speed of that Whisper-based app.

A test 4K 7GB video file was read and transcribed into subtitles by Apple Intelligence in 45 seconds. It took MacWhisper with the Large V3 Turbo LLM at total of 1 minute and 41 seconds.

Then the MacWhisper Large C2 model took 3 minutes and 55 seconds to do the same job.

None of these transcriptions were perfect, and all required editing. But the Apple Intelligence version was as accurate as the Whisper-based tools, and twice as fast.

As well as releasing these Apple Intelligence tools to developers, Apple has published videos with details of how to implement the technology.

The company has also, of course, leveraged the same technology across its beta releases of macOS 26, iOS 26, and more. It's the technology that drives its live translation in Messages and phone calls.



Read on AppleInsider

Comments

  • Reply 1 of 5
    9secondkox29secondkox2 Posts: 3,658member
    Finally a win for Apple Intelligence. 

    Hopefully it just goes up from here. 
    jas99appleinsideruserForumPost
     3Likes 0Dislikes 0Informatives
  • Reply 2 of 5
    Xedxed Posts: 3,258member
    Is this done completely on device?
     0Likes 0Dislikes 0Informatives
  • Reply 3 of 5
    thttht Posts: 6,010member
    Xed said:
    Is this done completely on device?
    Yes. Even MacWhisper is done on device. This is why time-to-finish here is a pretty big win.
    Xeddanox
     2Likes 0Dislikes 0Informatives
  • Reply 4 of 5
    Xedxed Posts: 3,258member
    tht said:
    Xed said:
    Is this done completely on device?
    Yes. Even MacWhisper is done on device. This is why time-to-finish here is a pretty big win.
    Then that it is wonderful news!
     0Likes 0Dislikes 0Informatives
  • Reply 5 of 5
    danoxdanox Posts: 3,879member
    I assume that the hardware used is the current Apple Silicon hardware, which means that when Apple releases the next iPhone and the next MacBook Pro in the fall of 2025, the software will/should be even faster. Very good news.

    And there appears to be some more good news and you begin to understand why Apple is not paying OpenAI one red cent.

    https://garymarcus.substack.com/p/a-knockout-blow-for-llms




    neoncat
     0Likes 1Dislike 0Informatives
Sign In or Register to comment.