Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
Apple IT

Apple M3 Pro Chip Has 25% Less Memory Bandwidth Than M1/M2 Pro (macrumors.com) 78

Apple's latest M3 Pro chip in the new 14-inch and 16-inch MacBook Pro has 25% less memory bandwidth than the M1 Pro and M2 Pro chips used in equivalent models from the two previous generations. From a report: Based on the latest 3-nanometer technology and featuring all-new GPU architecture, the M3 series of chips is said to represent the fastest and most power-efficient evolution of Apple silicon thus far. For example, the 14-inch and 16-inch MacBook Pro with M3 Pro chip is up to 40% faster than the 16-inch model with M1 Pro, according to Apple.

However, looking at Apple's own hardware specifications, the M3 Pro system on a chip (SoC) features 150GB/s memory bandwidth, compared to 200GB/s on the earlier M1 Pro and M2 Pro. As for the M3 Max, Apple says it is capable of "up to 400GB/s." This wording is because the less pricey scaled-down M3 Max with 14-core CPU and 30-core GPU has only 300GB/s of memory bandwidth, whereas the equivalent scaled-down M2 Max with 12-core CPU and 30-core GPU featured 400GB/s bandwidth, just like its more powerful 12-core CPU, 38-core GPU version.

Notably, Apple has also changed the core ratios of the higher-tier M3 Pro chip compared to its direct predecessor. The M3 Pro with 12-core CPU has 6 performance cores (versus 8 performance cores on the 12-core M2 Pro) and 6 efficiency cores (versus 4 efficiency cores on the 12-core M2 Pro), while the GPU has 18 cores (versus 19 on the equivalent M2 Pro chip).

This discussion has been archived. No new comments can be posted.

Apple M3 Pro Chip Has 25% Less Memory Bandwidth Than M1/M2 Pro

Comments Filter:
  • by Anonymous Coward

    But if the overall performance is still better than previous chips then who cares? Like all the morons here talking about needing more pcie lanes and then can't ever say they hit a bandwidth limit.

    • Re: (Score:2, Insightful)

      The onus is on you to define "better performance". We can discuss your point once we know what you actually mean.

      • by Anonymous Coward

        Is the M3 faster than every M1/M2 benchmark? This isn't complicated.

        • It is. Up to 40% means they probably found niche applications where such an improvement could be made. There might also be a milder general improvement in most tasks. A decrease in a certain spec could also mean performance regression in tasks that rely on those specs.
          • Possible, but unlikely.

            I have found it very difficult to hit 400GB/s on my M1 Max.
            It can do it, but only in pretty unrealistic workloads.
            That's not a fault of the chip, it's just the complication of using hardware bandwidths and assuming software will ever be able to achieve them.
            The CPU block (on the M1 Max anyway) has 200GB/s of bandwidth, and the GPU has about 300GB/s.
            If the GPU is unloaded, the CPU can hit 200 pretty easily, however when the GPU is loaded, the bottleneck on the CPU makes it fairly
      • Re:Ok (Score:5, Insightful)

        by MachineShedFred ( 621896 ) on Tuesday October 31, 2023 @01:10PM (#63969490) Journal

        I don't have a dog in this fight because I'm not looking to buy a new laptop from anyone in the next 18 to 24 months, but really this isn't hard. Consider the following postulated decision flow:

        1. M2 has X memory bandwidth, but usage statistics show it's almost never used because performance is bottlenecked elsewhere in the system. Thus, each part is more expensive because you're including capacity that is never used.
        2. M3 development possibly made two different prototypes due to #1; one with an equal X memory bandwidth, and one with X - n memory bandwidth. These two parts were then tested to see what the difference, if any, is.
        3. They discovered that only the highest core count M3 actually required X memory bandwidth, and chips with less active cores never got into using it, so having two parts with different memory bandwidth capacities makes sense.

        This is how this kind of thing usually comes about - the design has been refined and instead of just putting a massive memory controller and memory on regardless of cost, they've tuned it a bit better to reduce part and fab costs and complexity in order to deliver what's needed to make the whole system perform better, but not have unused expensive capacity unless it actually can be used.

        Really, it's no different than putting really fast RAM on your Ryzen CPU that doesn't get you any more performance than buying cheaper RAM that clocks 1:1 with the memory controller. Yes, technically there is more bandwidth there because the clock is higher, but it won't ever be used by the system during real operation.

        I don't know why people would be bitching about having a better tuned system, other than the fact that Apple will never pass on the cost savings being realized by having a better tuned configuration in the products.

        • The RAM comment isn't correct these days with AMD EXPO or Intel XMP.
        • by dfghjk ( 711126 )

          "I don't know why people would be bitching about having a better tuned system..."

          Are people bitching? It seems more a problem for Apple apologists who are losing a talking point. Most people wouldn't know or care.

          "Yes, technically there is more bandwidth there because the clock is higher, but it won't ever be used by the system during real operation."

          But it can be used by fanboys for the claims of superiority.

          • Apple users spend a lot less time worrying about justifying themselves than you think they do.

            • by Osgeld ( 1900440 )

              and I have never justified using a non apple product in my entire life

              • by mjwx ( 966435 )

                and I have never justified using a non apple product in my entire life

                To be fair, the only justifications you need are: It cost 1/3 the price and plays all the games a Macbook pro cant.

            • by mjwx ( 966435 )

              Apple users spend a lot less time worrying about justifying themselves than you think they do.

              Yeah, they only spend about 98% of their time doing it, rather than 99%.

        • by AmiMoJo ( 196126 )

          There main limitation in most M3 systems is going to be RAM. 8GB shared with the GPU is not a lot. My phone has more than that.

          • 8 GB even not shared with the GPU isn't a lot. My 13 year old (admittedly high end) laptop came with 16 (rocking 32 after a recent upgrade), and the obsolete PC I rescued from my in-laws after they replaced theirs had 6. It was a cheap desktop, about a decade old. It now has 16.

            8G is low end these days. Great on a, raspberry pi, pathetic on a premium machine.

            • At work the M1 with 8GB RAM I use outperforms the 2012 Mac Pro Trash Can we have with 128GB RAM and many more cores in the CPU (I forget the chip). Needless to say, both far outperform my 8GB RAM Raspberry Pi.

              I'm no EE so I'm not going to make claims about the technical details that make this the case, but from observation I know it's true. The best I can liken it to is that just because a BMW Z4 and a Honda Civic both have a 2.0 liter 4 cylinder engine doesn't mean you ought to expect the same performance.

              • Number of bytes of RAM is only very loosely related to performance in the general sense.
                So I don't really know how to parse what you said.

                Your 8GB Air will absolutely perform worse than the Mac Pro Trash Can if you try to do something that has a 16GB working set.
                My M1 MBA has 16, my coworker's has 8. He came to regret the 8 pretty quickly, since it's pretty easy to end up with ~20GB of actively used memory in macOS with a busy desktop, turning your poor air into a swapping nightmare.

                Even 16 ended up b
              • If you're doing work which requires less than 8G then why on earth did you buy a 128G machine a decade ago? That must have cost a fortune.

                RAM doesn't help performance until you don't have enough of it. If your data in ram approaches the limit of memory your performance will tank. Before that though, RAM is faster than any SSD so file caches or ram disks can hugely boost performance.

            • 8 GB even not shared with the GPU isn't a lot. My 13 year old (admittedly high end) laptop came with 16 (rocking 32 after a recent upgrade), and the obsolete PC I rescued from my in-laws after they replaced theirs had 6. It was a cheap desktop, about a decade old. It now has 16.

              8G is low end these days. Great on a, raspberry pi, pathetic on a premium machine.

              Memory requirement generalities on one Platform have little to do with those on another Platform.

              macOS just seems to do better (much better!) than, say Windows. It's just a fact.

              And people buying a Budget laptop are probably not editing 8 streams of 4k Video as part of their Job.

              • Memory requirement generalities on one Platform have little to do with those on another Platform.

                That's just not true. The memory floor varies from platform to platform, but a browser will be the same level of hoggyness on any platform.

                macOS just seems to do better (much better!) than, say Windows. It's just a fact.

                Lol. Sure, but I'm comparing Linux and OSX, not Windows and OSX.

                And people buying a Budget laptop are probably not editing 8 streams of 4k Video as part of their Job.

                They're also not buying a mac

                • Memory requirement generalities on one Platform have little to do with those on another Platform.

                  That's just not true. The memory floor varies from platform to platform, but a browser will be the same level of hoggyness on any platform.

                  macOS just seems to do better (much better!) than, say Windows. It's just a fact.

                  Lol. Sure, but I'm comparing Linux and OSX, not Windows and OSX.

                  And people buying a Budget laptop are probably not editing 8 streams of 4k Video as part of their Job.

                  They're also not buying a mac since it's not a budget laptop. It's the brand that gives you budget specs for premium prices.

                  1. You never mentioned Linux.

                  2. Why do you bring up Browsers?

                  3. Your In-Laws ran Linux? Sure. . .

                  4. Budget Specs? Not what reviewers and owners are saying about Apple Silicon Macs.

                  • 1. I never mentioned windows. Why did you assume.

                    2. They are memory hogs and everyone used one

                    3. Eh?

                    4. Yeah budget specs. 8G is very much a budget machine level of ram.

        • I took note of the small die size for the base M3 chip. That's the one they are going to ship the most of. That one needs to have the lowest cost to maintain margins in all the products that have it (laptops, mini, iPad pro, etc).

      • System performance is a multifaceted evaluation, to which there is no single lynchpin that guarantees better overall operation, and designers make trade-offs based on most common use cases.
    • by dfghjk ( 711126 )

      You do, considering you're posting about it. Worried?

    • by slaker ( 53818 )

      My workstation has video cards with nVidia and Intel chipsets, a QDR Infiniband HBA, a 4xu.2 HBA and a couple extra 1xm.2 to PCIe adapters. I don't think I've ever hit the LIMIT of any of the PCIe 4.0 connectors on my motherboard, but I'm certainly happy that I have enough lanes to have that much I/O available to me.

      (The Intel GPU is present because it offers hardware video compression support that nVidia and AMD don't provide; Infiniband is far and away the cheapest way to get a fast connection to my file

      • by CAIMLAS ( 41445 )

        Why would you care that the board has that much bandwidth available to you if a) you're not using it, b) you can't possibly use it?

        That's the scenario here. This isn't a nerf, it's an architectural optimization. There is next to zero chance they did something so stupid as to decrease performance by decreasing memory bandwidth, after all the time and effort they put into chip development.

        • Why would you care that the board has that much bandwidth available to you if a) you're not using it, b) you can't possibly use it?

          That's the scenario here. This isn't a nerf, it's an architectural optimization. There is next to zero chance they did something so stupid as to decrease performance by decreasing memory bandwidth, after all the time and effort they put into chip development.

          This.

      • accessing memory over PCIe is very slow and high latency. You likely don't hit bandwidth limits because there is a natural tendency for high bandwidth devices like GPUs to avoid accesses that cross between CPU's memory controller and the PCIe bus.

    • You're the moron. Here is a fine example of what happens when you don't have enough PCIe bandwidth. The card loses about 8-10% of its performance when going from PCIe Gen 4 to Gen 3.

      https://www.tomshardware.com/r... [tomshardware.com]

      • The point was not there is no performance loss if there is PCIe bandwidth throttling. The point is that some people complaining about [System A] not having the same PCIe bandwidth as [System B] would never use enough PCIe bandwidth for that difference to matter. For example the NVidia fans crowing their RTX 3070 cards have PCIe 4x 16 over Radeon 6600XT owners who only had PCIe 4 X 8. Neither card use enough of the PCIe bandwidth where that mattered.
      • Wow, whoosh.
        That is an example of what happens when you limit the bandwidth of a system that is using that bandwidth. Yes, its performance drops.
        However, if you limit the bandwidth on a system that does not use it, there will be no performance cost.

        This isn't rocket science. Try not to look like so much of a moron while calling people a moron in the future.
        • Wow, whoosh.

          That is an example of what happens when you limit the bandwidth of a system that is using that bandwidth. Yes, its performance drops.

          However, if you limit the bandwidth on a system that does not use it, there will be no performance cost.

          This isn't rocket science. Try not to look like so much of a moron while calling people a moron in the future.

          Exactly!

          For an alleged tech-site, Slashdot sure seems to have a dearth of meatsacks that have even a modicum of actual Engineering savvy. . .

    • Exactly.
    • M3 is one better than M2, so of course the performance is better! /sarcasm

  • by sapgau ( 413511 ) on Tuesday October 31, 2023 @12:18PM (#63969376) Journal
    Some paragraphs down it says:
    "Taken together, it's presently unclear what real-world difference these changes make to M3 performance when pitted against Apple's equivalent precursor chips in various usage scenarios, especially given that the latest processors include new Dynamic Caching memory allocation technology which ensures that only the exact amount of memory needed is used for each task. "

    Maybe that Dynamic Caching makes it hard to compare and probably:
    - The performance might be very close
    - You might not notice any difference
  • Binning? (Score:4, Interesting)

    by gotamd ( 903351 ) on Tuesday October 31, 2023 @12:29PM (#63969398)
    I saw a Youtube video from Snazzy Labs that discussed these points and speculated that it's due to yields from the TSMC 3NB process, which is presumably what these chips are produced on (like the A17 Pro in the new iPhone Pro 15). I'm guessing that due to poor yields, Apple finds themselves having to disable cores in quite a few of the M3-based chips, which is why we see decreased memory bandwidth and core count along with so many different variations of the chip vs. in prior generations. Does that have a performance impact? I'm sure it does. Does it make the chips perform worse than the M1 and/or M2 in any real-world benchmarks? I guess we'll find out soon.

    I wonder if we'll see an M3 Macbook Air at some point. Sticking the M3 only in the iMac and 14" Pro will ensure there isn't a huge demand for it and it will also keep margins higher. Maybe the MBA will skip the M3 generation altogether, or maybe if yields of Pro and Max binned chips are bad enough and demand for the iMac and 14" MBP w/ vanilla M3 are low enough, Apple will throw an M3 refresh MBA out there.

    The only thing about the M3 that currently looks appealing is AV1 decode support. My guess is that the M4 generation, still on "3nm" but maybe a better yielding version, will probably "fix" a lot of these oddities with core count and memory bandwidth on the M3 generation.
    • by wwphx ( 225607 )
      Kinda makes me glad I got a 14" M2 MBP earlier this year. I'm extremely happy with the performance, not that I expect to max it. I'm ecstatic about the form factor: much smaller and lighter than my 15" that it replaces and easier to grip as my hands get older. Even though the 16" is the same size as my 15" 2015 MBP, I find the 14" screen to be extremely comfortable along with the slightly smaller keyboard.
      • I got the 16". My CEO has the 14".

        Form factor wise, I am envious of his.
        But sometimes I do find myself happy that I got mine. The big screen is nice. It's just not very portable.
        • I got the 16". My CEO has the 14".

          Form factor wise, I am envious of his.

          But sometimes I do find myself happy that I got mine. The big screen is nice. It's just not very portable.

          It's a bigger screen in virtually the same size machine as yesteryear's 15" MBP.

          What's not to like?

          • It's a bigger screen in virtually the same size machine as yesteryear's 15" MBP. What's not to like?

            Depends on how much lifting the word 'virtually' is doing there, really.

            • It's a bigger screen in virtually the same size machine as yesteryear's 15" MBP.

              What's not to like?

              Depends on how much lifting the word 'virtually' is doing there, really.

              This comparison is from 2019; but that 16" is close to the current MBP 16".

              14.09" x 9.68" for 2019 16" MBP. 14.01" x 9.77" for the 2023 16" MBP. And the 2019 15" MBP was 13.75" x 9.48".

              As I said, virtually the same size.

              https://9to5mac.com/2019/11/14... [9to5mac.com]

              https://www.apple.com/macbook-... [apple.com]

          • I thought that was pretty clear- it's not very portable.
            It's big, and it's heavy. I still use my air for field work.
            • I thought that was pretty clear- it's not very portable.

              It's big, and it's heavy. I still use my air for field work.

              Whatever.

              • Shrug, you asked.
                The smaller machine is infinitely more portable.
                I won't deny that I don't love the big screen on the 16" when I'm not lugging it around, though.
    • by CAIMLAS ( 41445 )

      I suspect Apple with settle into some sort of tick-tock-boom cadence with chip releases similar to this:

      * introduce a new hardware platform at the low end for the air, etc.
      * release the 'tock', a linear improvement from the original (eg. the M2)
      * release the 'boom' - aka the "maximum capability" for the platform

      Run that out for 6-8 years, then introduce a new architecture and start over...

    • I saw a Youtube video from Snazzy Labs that discussed these points and speculated that it's due to yields from the TSMC 3NB process, which is presumably what these chips are produced on (like the A17 Pro in the new iPhone Pro 15). I'm guessing that due to poor yields, Apple finds themselves having to disable cores in quite a few of the M3-based chips, which is why we see decreased memory bandwidth and core count along with so many different variations of the chip vs. in prior generations. Does that have a performance impact? I'm sure it does. Does it make the chips perform worse than the M1 and/or M2 in any real-world benchmarks? I guess we'll find out soon.

      I wonder if we'll see an M3 Macbook Air at some point. Sticking the M3 only in the iMac and 14" Pro will ensure there isn't a huge demand for it and it will also keep margins higher. Maybe the MBA will skip the M3 generation altogether, or maybe if yields of Pro and Max binned chips are bad enough and demand for the iMac and 14" MBP w/ vanilla M3 are low enough, Apple will throw an M3 refresh MBA out there.

      The only thing about the M3 that currently looks appealing is AV1 decode support. My guess is that the M4 generation, still on "3nm" but maybe a better yielding version, will probably "fix" a lot of these oddities with core count and memory bandwidth on the M3 generation.

      WTF, over?!?

      There are no more variants of the M3 than the M2 (actually less, since the M3 Ultra has yet to drop).

      There is no M3 Air because Apple is either burning through some inventory of M2 MBAs, and wanted to avoid The Osbourne Effect; or they wanted to showcase the MacBook Pros, which reportedly suffered recently from slower sales, before everyone jumping on an M3-powered lower-margin-but-still-more-than-enough-computer-for-most-people MacBook Air.

  • by laughingskeptic ( 1004414 ) on Tuesday October 31, 2023 @12:39PM (#63969418)
    Independent testing by PC Mag on the M1 Pro couldn't see a different between low power and high power modes ... https://www.pcmag.com/news/tes... [pcmag.com] .
    • by guruevi ( 827432 ) on Tuesday October 31, 2023 @12:57PM (#63969456)

      Depends on what exactly you're measuring and whether the system has enough power on both ends of the bus to use the bandwidth in the first place. It's perfectly reasonable to think that Apple may have over-provisioned in their first iterations and after testing found the cores or memory chips can't keep up anyway. Or found other ways to boost the performance.

      We see the same issue with DDR5 and PCIe5.0 in PC, if all you're doing is putting a GeForce GT 1080 chip on a PCIe5.0 bus, it wouldn't matter if the manufacturer takes out half the lanes, which is what you often see on budget motherboards (x16 slots operating at x4 or x8)

      • by dfghjk ( 711126 )

        Except that Apple specifically added this feature, presumably because it did actually make a difference.

        "Or found other ways to boost the performance."

        You seem confused. Apple found THIS way to "boost the performance".

  • Apple chips are basically the most easily accessible, high-performance hardware for at-home/open-source LLM usage. Memory bandwidth and size are the key constraints on inference speed. Maybe it's intended, maybe it's a happy coincidence for Apple that anyone who wants the best LLM performance on a Mac is going to be pushed into buying the most expensive tier(s). I suspect they knew what they were doing.

    • by dfghjk ( 711126 )

      For all those masses implementing their own LLMs at home on consumer hardware. I mean, who doesn't keep their own personal few hundred terabytes of training data around to push though their MacBook Pro?

      • For all those masses implementing their own LLMs at home on consumer hardware. I mean, who doesn't keep their own personal few hundred terabytes of training data around to push though their MacBook Pro?

        Hundreds of terabytes? The pile is less than a TB... World class models are trained on something on the order of 4T tokens. Besides I'm pretty sure everyone is talking about inference rather than training.

        I only have about 150GB of memory bandwidth on my workstation so even taking full advantage of AMX tile registers I'm lucky to see a token a second (non batched) on something like a Falcon 180B model perhaps closer to 2 if heavily quantized. With a MAC you can get over twice that which for the price and

    • This is exactly why I am not buying an M3 MBP. I already run the 34B parameter version of Code Llama on my laptop, and it's all limited by memory bandwidth. The M3 doesn't improve over the last generation 400GB/s, so there is no speedup (aside maybe for prompt processing). Apple can charge $4000 (or whatever they want) to run these workloads, since other laptops can barely do it.
  • by Petersko ( 564140 ) on Tuesday October 31, 2023 @03:08PM (#63969786)

    Apple haters spend an awful lot of time imagining things that apple fans must be saying to justify themselves and refuting them.

    Apple users spend little to no time worrying about justifying themselves.

    It's a debate where one side is frothing at the mouth, and the chairs on the other side of the stage are empty.

    • Indeed this reminds me of console fanboyism where they think game software company workers are part of their team. People in the gaming industry are not as die hard as fanboys think. Game developers who work on Xbox play games on Switches and *gasp* PS5s.
    • I think some of it goes back decades and decades and at one point there were interesting or nuanced arguments to be had, and in some cases I think there still are, but there are always some people for whom the war never really ended and they're still fighting battles that a lot of people weren't born in time to really understand.

      There're also just trolls looking to get a reaction. Being unhinged and mouth-frothy is a good way to get a reaction. A well-reasoned and mild-mannered comment can be consumed an
    • This is funny. It sounds to me like you're spending a bit too much time thinking about what these Apple "haters" are thinking about.
    • Apple haters spend an awful lot of time imagining things that apple fans must be saying to justify themselves and refuting them.

      Apple users spend little to no time worrying about justifying themselves.

      It's a debate where one side is frothing at the mouth, and the chairs on the other side of the stage are empty.

      Well said, bravo!

    • by mjwx ( 966435 )

      Apple haters spend an awful lot of time imagining things that apple fans must be saying to justify themselves and refuting them.

      Apple users spend little to no time worrying about justifying themselves.

      It's a debate where one side is frothing at the mouth, and the chairs on the other side of the stage are empty.

      Irony is not your strong suit.

      You've made at least two posts on this article saying the same thing.

      You seem to spend an inordinate amount of time trying to justify why you've bought Apple products and trying to dismiss anyone who disagrees with you as "haters".

      You did get one thing right, only one side of this argument is "frothing".

  • "Apple M3 Pro Chip Has 25% Less Memory Bandwidth Than M1/M2 Pro"

    It's good to see Apple conserving bandwidth like this, I bet Lenovo and Dell wished they'd thought of it. Brilliant.

As you will see, I told them, in no uncertain terms, to see Figure one. -- Dave "First Strike" Pare

Working...