M5 Pro may separate out GPU and CPU for new server-grade performance
Analyst Ming-Chi Kuo says that Apple will move away from its current processor designs that keep the CPU and GPU cores on the same chip -- and see a performance gain.

TSMC has just announced an all-new chip production process called "A16"
One of the reasons for Apple Silicon's speed over the previous Intel processors has been that each M-series chip has been a single unit. This System-on-a-Chip (SoC) idea cuts bottlenecks by having all the processor's elements together on one chip package.
According to Kuo, however, Apple is going to change this for the M5 Pro, M5 Max, and M5 Ultra. Only the M5 will remain as a single unit.
Instead, the M5 Pro and other chips will use manufacturer TSMC's latest chip packaging process. Called the System-in-Integrated-Chips-Molding-Horizontal (SoIC-mH), it puts together different chips into one package.
Apple M5 series chip
1. The M5 series chips will adopt TSMC's advanced N3P node, which entered the prototype phase a few months ago. M5, M5 Pro/Max, and M5 Ultra mass production is expected in 1H25, 2H25, and 2026, respectively.
2. The M5 Pro, Max, and Ultra will utilize https://t.co/XIWHx5B2Cy-- (Ming-Chi Kuo) (@mingchikuo)
The advantage, according to Kuo, is that this will produce "server-grade" packaging. Apple "will use 2.5D packaging" that has "separate CPU and GPU designs," and which will "improve production yields and thermal performance."
Kuo says that mass production is expected in 2H25 for the M5 Pro and the M5 Max, and then 2026 for the M5 Ultra. The M5 has reportedly been in the prototyping phrase for a few months, and mass production is believed to be planned for 1H25.
That M5 processor will be produced by TSMC using its N3P technology, which is expected to be seen first in the iPhone 18 range.
Kuo also claims that the M5 Pro processors will be used in Apple Intelligence servers. Specifically, it will be utilized for the company's Private Cloud Compute technology.
Rumor Score: Possible
Read on AppleInsider
Comments
Oh, and separating the memory from CPU/GPU appears to be a step backwards to UMA, whatever AMD is pursuing right now it isn’t better than Apple Silicon when it comes to wattage used efficiency and performance, yes they (AMD) are closer than Intel but Nvidia is just completely out of the ballpark when it comes to energy wasted, yes it’s faster for certain tasks now but 1000-1500 watts liquid cooled systems are an eventual dead end which Apple used at one time when they were at the heights of their G5 hell.
https://www.digitaltrends.com/computing/why-nvidia-rtx-4090s-are-melting/ butterfly keyboards, antenna gate, bend gate, power button location, or upside down mice are nothing to this….. I don’t think Apple will go away from power efficiency.
I can think of 3 reasons this would make sense:
1. If Apple is combining their silicon with third party chips from Nvidia or semthing.
2. Alternatively, another way it could make sense is if Apple is looking to add more GPU cores to various iterations of its chips, without increasign CPU core counts. i.e. having multiple sets of max/Ultra chips - one set for laptops, one for Mac Studio, and another for Mac Pro.
or...
3. Appel could be redoing the way it tiers its chip lineup. CPU could be the same, but GPU would be different for each tier.
Interesting to see how this develops.
To add more resources the only thing you can do is manufacture multiple chiplets each themselves at the reticle limit and package them together.
So basically, like HBM but with logic chips or mix and matching? There’s going to be chip scale heat transfer plates sandwiched between layers eventually.
Apple will have to go away from power efficiency if they want higher performance, unless we change how physics work. And connecting multiple macs together has been there for a while with MLX, and it works fantastic.
I hope these rumors don’t mean Apple is not going to the release the M4 Studio Ultra in 2025 if so, that means they won’t be releasing any Studio Ultra until the middle to end of 2026 which would lead to further collapse in Mac desktop sales at the wrong time when many people were starting to look at big memory high bandwidth Macs to run AI models locally.
https://odysseyapp.io/blog/five-reasons-to-run-ai-models-locally
https://forums.macrumors.com/threads/so-happy-with-the-m4-pro-i-can-finally-use-ai-stuff-locally.2442964/
None of which means that consumer machines wouldn’t still be single chip SoCs.