Heavily upgraded M3 Ultra Mac Studio is great for AI projects

Jump to First Reply
Posted:
in Mac Software

The Mac Studio is a great system for running AI models like DeepSeek locally. That is, if you're prepared to pay for M3 Ultra and a lot of upgrades.

Silver rectangular device with ports, situated on a desk. Nearby are a small red speaker and a computer monitor displaying vibrant colors.
Mac Studio



Apple Silicon is dominating the AI-capable PC market, with its capability for machine learning applications making it an attractive purchase. In a new video, it seems that the M3 Ultra Mac Studio could offer people working in AI a considerable amount of performance to play with.

The video from Dave2D discusses the M3 Ultra Mac Studio, Apple's fastest Mac on the roster. However, what is being tested in the video is more to demonstrate the extremes of the hardware, rather than a more modest and conventional loadout.

The version shown in the video uses an M3 Ultra chip with the upper-tier configuration of a 32-core CPU, an 80-core GPU, and the 32-core Neural Engine. It's also packed with 512GB of unified memory, the maximum amount available for the model, with a memory bandwidth of 819GB/s.

Local LLM usage



While the review discusses how the model doesn't make a massive change to typical video content creator workflows, it does focus on the massive amount of memory. That, and its capability to run massive Large Language Models (LLMs) used for AI applications.

The most obvious use is to be able to use an LLM locally without needing to send the request out to a server farm. In cases such as a hospital, the need for patient privacy means keeping the data on-site is a better option than sending it off-site to be processed, where possible.



To test it out, Deepseek-r1 was loaded and ran locally on the Mac Studio. This wasn't easily possible before, due to the 671-billion-parameter model requiring just over 400 gigabytes of storage and a bit less than 450 gigabytes of video RAM to function.

Since Apple Silicon uses unified memory, the top 512GB memory configuration is able to handle the massive file size of the model. Though you can run lower-parameter versions on a Mac with a smaller amount of memory, you can only really use the highest configuration to use the biggest Deepseek-r1 model.

Indeed, during testing, it was found that macOS limits the amount of memory that can be used for video memory by default to 384GB, so that had to be overridden for testing to begin.

When it was up and running, the Mac Studio was able to churn through queries at approximately 17 to 18 tokens per second. This is a level that is considered usable for the majority of LLM uses.

While LLM performance is one factor of interest, another is power consumption, as running the models can require a lot of resources. In the case of the model on the Mac Studio, it was observed as requiring 160 to 180 Watts during use.

This may sound like a lot, but it is relatively small when compared against a custom-built PC with multiple GPUs made for the same task. It's proposed that the power draw of that hypothetical system could be ten times that of the Mac Studio.

An expensive option



Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

The 512GB option is only available for the upper-tier M3 Ultra Chip, which adds $1,500 to the base model's $3,999 price. Going from the base 96GB memory to 512GB is a $4,000 add-on again, bringing the total cost to $9,499 with the base 1TB of storage untouched.

Dropping around $10,000 on a Mac for AI purposes is out of the question for most people. But, it can be an option for businesses and corporations who have the finances and can more easily justify having a Mac Studio with M4 Ultra to run an on-site LLM.

Such operations may even consider creating a cluster of Macs, if they have the budget.



Read on AppleInsider

«1

Comments

  • Reply 1 of 38
    blastdoorblastdoor Posts: 3,736member
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    williamlondoniooineoncatAlex1Nbaconstangwatto_cobra
     5Likes 1Dislike 0Informatives
  • Reply 2 of 38
    RDWrdw Posts: 8member
    Well, DUH! Who knew if you throw a lot of money and spec out a machine to the fullest with tons of RAM and GPU and CPU cores, it would be a beast of a machine?
    williamlondonwatto_cobra
     2Likes 0Dislikes 0Informatives
  • Reply 3 of 38
    CarmBcarmb Posts: 103member
    Considering how much is being made regarding the power needed to implement AI, the most important element of Apple's hardware is its efficiency. Could be a dramatic advantage going forward. 
    williamlondoniooineoncatAlex1Nbaconstangdavwatto_cobra
     6Likes 1Dislike 0Informatives
  • Reply 4 of 38
    Why on earth would I want to run an AI model?  Locally or otherwise?
    blastdoorargonautwilliamlondoniooiITGUYINSDneoncat9secondkox2davwatto_cobra
     4Likes 5Dislikes 0Informatives
  • Reply 5 of 38
    brianusbrianus Posts: 184member
    Why on earth would I want to run an AI model?  Locally or otherwise?
    I’m sure this was meant to be snarky, but for me it’s a genuine question: what are the envisioned real world use cases? What might a business (even a home one) use a local LLM for?

    The article mentions a hospital in the context of patient privacy, but what would that model actually be *doing*?
    ITGUYINSDneoncatAlex1N9secondkox2davwatto_cobra
     5Likes 1Dislike 0Informatives
  • Reply 6 of 38
    blastdoorblastdoor Posts: 3,736member
    brianus said:
    Why on earth would I want to run an AI model?  Locally or otherwise?
    I’m sure this was meant to be snarky, but for me it’s a genuine question: what are the envisioned real world use cases? What might a business (even a home one) use a local LLM for?

    The article mentions a hospital in the context of patient privacy, but what would that model actually be *doing*?
    Anybody who still doesn’t see the value of AI does not have a bright future career wise. But maybe the people still making these comments are retired, in which case I’m jealous — I’ve got a lot of years left.
    danoxargonautwilliamlondoniooiITGUYINSDneoncat9secondkox2davwatto_cobra
     6Likes 3Dislikes 0Informatives
  • Reply 7 of 38
    Why on earth would I want to run an AI model?  Locally or otherwise?
    If you go to ChatGPT and ask "Do people have different computing needs?" You will have successfully used an LLM to answer your own questions.
    danoxiooiAlex1N9secondkox2
     3Likes 1Dislike 0Informatives
  • Reply 8 of 38

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    williamlondonAlex1Nbaconstangwatto_cobra
     4Likes 0Dislikes 0Informatives
  • Reply 9 of 38
    blastdoorblastdoor Posts: 3,736member

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    watto_cobra
     1Like 0Dislikes 0Informatives
  • Reply 10 of 38
    ApplePoorapplepoor Posts: 363member
    Folks, this is an Apple device and one pays dearly to play in the Apple world. 

    Back in the early 1990s, Apple's IIfx was king of the Apple computer mountain. Fully loaded with 32Mb of third party memory, a very heavy 21" Radius color CRT monitor with Radius video card and a 80 MB hard drive, the cost was close to $12,000 in 1990 dollars. I checked and that amount in today's dollars is over $29,000 so the current loaded Mac M3 Ultra Mac Studio at just over $14,000 is a relative bargain.

    Photoshop v1.3 was on several 3.5" discs. 

    One buys the tool for the job. If one is not doing the "AI" thing, then the M4 Max Mac Studio is probably sufficient or even the M4 Pro mini which is close to the power of the M1 Ultra Studio with nearly double the single core speed of the M1 Ultra.
    edited March 18
    argonautwilliamlondonneoncatAlex1Nbaconstangwatto_cobra
     4Likes 2Dislikes 0Informatives
  • Reply 11 of 38
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    9secondkox2baconstangwatto_cobra
     2Likes 1Dislike 0Informatives
  • Reply 12 of 38
    tobiantobian Posts: 158member
    With traditional PC setup, 10 times more power draw means approx. 10 times more heat to dissipate. One would imagine some huge active cooling box and alot of cables.. but there sits one little great Mac Studio on the desk. Very impressive!!
    argonautwilliamlondondanoxneoncatAlex1Nbaconstangwatto_cobra
     6Likes 1Dislike 0Informatives
  • Reply 13 of 38
    blastdoorblastdoor Posts: 3,736member
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 


    williamlondonbaconstangwatto_cobra
     1Like 2Dislikes 0Informatives
  • Reply 14 of 38
    DOGE proposes replacing the current heat generating Oval Office “small language” processor with this cost-effective large language Apple Mac Studio. Estimated first-year savings, over $1 trillion. Downsizing required? Just one person.
    ITGUYINSDAlex1Ncpsro9secondkox2baconstangrezwitswatto_cobra
     5Likes 1Dislike 1Informative
  • Reply 15 of 38
    sunman42sunman42 Posts: 331member
    blastdoor said:
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 


    Well, one not too shabby version ran on my *cough* poor, old (yeah, 18 month old) M2 Ultra Studio with a mere 128 Gbyte of memory. And yet, it was still able to give reasonable answers to two of the burning questions of the last ~ 50 years: (1) What is the air speed of an unencumbered swallow? and (2) Lakes or hamantashen? It appeared to be familiar with both source texts and more recent work online. Curiosity satisfied, and not quite certain what, if anything, it might have been communicating to a mother ship somewhere, I deleted DeepSeek-r1:<whatever>. Who knows, might be the last time I ever mean to ask an AI anything.
    williamlondonwatto_cobra
     1Like 1Dislike 0Informatives
  • Reply 16 of 38
    ITGUYINSDitguyinsd Posts: 565member
    DOGE proposes replacing the current heat generating Oval Office “small language” processor with this cost-effective large language Apple Mac Studio. Estimated first-year savings, over $1 trillion. Downsizing required? Just one person.
    It couldn't do any worse than the one bit "processor" we have there now.
    cpsrobaconstangwatto_cobra
     3Likes 0Dislikes 0Informatives
  • Reply 17 of 38
    cpsrocpsro Posts: 3,266member
    What's the largest, full precision Deepseek r-1 model that will run on a 512GB Studio? And how does it compare in speed and results to the reduced precision 670b parameter Q4-K-M model?
    williamlondonwatto_cobra
     1Like 1Dislike 0Informatives
  • Reply 18 of 38
    blastdoor said:
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 



    blastdoor said:
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 


    So… when I said “this is covered in the video” I literally meant that is was covered in the video. All we have established here is that  you haven’t watched the video. 
    neoncatblastdoorwilliamlondonbaconstangbulk001watto_cobra
     3Likes 2Dislikes 1Informative
  • Reply 19 of 38
    blastdoorblastdoor Posts: 3,736member
    blastdoor said:
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 



    blastdoor said:
    blastdoor said:

    blastdoor said:
    Even though the Mac Studio with M3 Ultra seems like a great option for LLM usage and development, there is a big drawback in terms of cost.

    But compared to what? How much would you have to spend to do the same job on a PC?
    This was actually covered in the video. You can rent a ton of server time for the cost of a maxed out Mac Studio. So if you are just using it for development or general LLM usage then it really doesn't make sense financially. That said, the maker of the video also gave examples of why it would be worth it to pay to run an LLM locally. Specially when it has to do with privacy.
    My question isn’t the cost to rent but the cost to buy.
    If that was the intent of your questions then your questions were really poorly worded as buying didn't come up at all.

    Anyway, to answer your question, you could build a PC that could to this cheeper than you could buy a Mac Studio. The big deal about Deepseek was that it ran on consumer hardware. 
    There are many versions and the 671 billion parameter version is not going to run on anything resembling a standard PC. So I think you just don't know. 


    So… when I said “this is covered in the video” I literally meant that is was covered in the video. All we have established here is that  you haven’t watched the video. 
    At about 5 minutes and 30 seconds he says that building this with consumer PC hardware would be "quite expensive." I was looking for a fair bit more precision than that. 

    williamlondonmuthuk_vanalingamwatto_cobra
     2Likes 1Dislike 0Informatives
  • Reply 20 of 38
    Rogue01rogue01 Posts: 241member
    There is an error at the end of the article.  M3 Ultra Mac Studio, not an M4 Ultra Mac Studio.

    The M3 Ultra is only good for those AI models.  Other benchmarks confirm it is slower than the M4 Max.  And of course Apple has to rip people off with $4,000 for 512GB of memory.
    watto_cobra
     0Likes 1Dislike 0Informatives
Sign In or Register to comment.