Nuvia: don’t hold your breath

At MR, yesterday someone posted specifics they pulled from Geekerwan's video. And a bit before that they posted block diagrams of the new cores. They strongly suggest that this is a whole new ball game!

We are looking at not one, but TWO new cores. They are calling them "L" and "M", suggesting that perhaps an "S" might come along at some point. The M cores are a LOT smaller - less than 50% the size - with 4-wide decode instead of 8, half (sort of) the execution resources, smaller queues and register files, maybe smaller branch predictor, etc.

I think QC has come a lot closer than we thought they were going to manage to Apple. If they manage to add SME in the next get they will likely be neck-and-neck with the M4, and if they keep pushing on the rest at the same time they could plausibly match whatever is in the M5. That's a really big deal! If all these numbers hold up, we finally have a real horse race.
:)
 
I think people here were underestimating how competent these guys are just because Apple is still good, while skepticism of Qualcomm is way higher than deserved — AMD/Intel are the laggards and have been for some time. Now sure, Qualcomm can act as a brick wall against Nuvia and maybe they’re too artisanal or whatever for them, but by all evidence they are not and did not want to, whereas with Intel internal fights and catabolism is just Tuesday and reportedly part of why they didn’t sell to Intel (don’t quote me tho).

But to some degree as I pointed out, their obstacles were one-time, the lawsuit, or getting in the groove of things with the X Elite which is probably why power wasn’t as good as we expected and there are multiple rumors and speculation on that. The fact that the turnaround was so significant and catapulted them ahead of the X925 for now is big.

It’s a real ball game for now.

Things do take time, so holding your breath from 2021 wasn’t wise and Cliff’s skepticism was warranted for sure, and the initial product was disappointing for the time (just less so relative to the cost and scalability imo). So in a way he was right :).

But from the point we saw they were shipping product with a reasonable advantage on Intel/AMD and then putting them in *phones* right afterwards — that was a huge tell to me. They couldn’t put something in that was objectively worse or at least not a little better and we knew where the X925 and such were going.
 
I also have a little essay from Reddit a few of you might like on area and Lunar Lake vs the M2 (not even M3, it’s more M2 comparable in some ways), and QC vs Strix Point.
This actually doesn’t explain as much as you would think.

Lunar Lake on N3B is 139mm2 for the main compute die, and a 4c Performance core CPU complex (including the L3 as this is important for these cores in a similar way Apple’s big shared L2 is) is around 26mm^2 for cores that, in Lunar Lake, are around 4.5-5.1GHz and M2 ST or M2 ST performance + 5-10% at best. And at 2-4x more power.

Do you know what a 4 performance core cluster is on an M2? It’s about 20.8mm2 on N5P.Yes, that includes the big fat L2.

Intel also has a big combined L1/0 now and 2.5MB private L2 for each P core, totaling 10 MB of L2, and 8 or 12MB of L3 depending on the SKU, though the area hit from 12MB will there either way (the marginal 4 is fused off.). In total for a cluster Intel is using 10MB of L2, 12MB of L3, vs 16MB of L2 with Apple.***

So Intel is using not only literally more core and total cluster area, but also just more total cache for a cluster of 4 P cores, and doing so on N3B vs N5P with a result that is at best 10% or so better in ST at 2-3x the power, and modally from reviews maybe 5% better on ST and again, much worse efficiency. And that’s just the M2.

It’s really just not true they’re (even AMD) notably better with CPU area specifically. It looks even worse if you control on wattages — because getting more “performance” by ballooning cores and cache for an extra 20% frequency headroom at massive increases in power is the AMD/Intel way, except this isn’t really worth it in laptops.

***And Apple has an 8MB SLC, that’s about 6.8mm2 but so do Intel on Lunar Lake at a similar size. Not a huge deal for area and similar for both.

—- Part II, AMD V Qualcomm N4P

We see this also in Qualcomm vs AMD. A single Oryon 4C cluster with 12MB of L2 is ~ 16mm2 on N4P and blows AMD out on ST performance/W (and only reason MT perf/W suffers is when QC is pushed too hard by default settings, it is still quite efficient dialed down), while still competing with Lunar Lake pretty well despite Lunar’s extra cache and other advantages.

By contrast, AMD’s 4 Zen 5 cores with their 16MB L3 are about 27mm^2, and the ST advantage you get is about 10-20% over the 3.4GHz standard Oryon (which not all SKUs will be anyway) albeit at 5-15W more power and with a crippling performance L at 5-12W vs QC. Not worth it.

The 8 Zen 5c cores with 8MB L3 are 30-31mm^2, which isn’t bad, except those have a clock hit to around ~ 4GHz and are even less efficient than regular Zen 5 at those frequencies both due to the design and the 1/4 the L3 per core. So, also not great.

It’s hard not to conclude Apple and yes Qualcomm and likely Arm too, are just winning on plain design & tradeoffs. — Because they are.
 
Also some CB24 MT data from the wall with the display off, just showing how solid the area and wattage efficiency can be here.

On the Qualcomm MT thing, here is CB2024 from the wall with an external monitor going: notice that Qualcomm can get top notch performance in a good power profile and efficiency, we just don’t know what they look like below 30W or so — would efficiency improve or decline? But either way at 35-45W these things are decent and nearly as good as they are at 60-100, and even beat AMD’s stuff at these wattages. Note this is from the wall, though might not be minus idle so it’s possible the others like AMD especially would do better with that.

Either way it’s not bad, but what is bad is people bullshitting about Qualcomm efficiency by implying it needs the 70-100W guzzler figures we’ve seen for some cases at wall power or for motherboard. Yes the peak figures are insane tradeoffs and OEMs are dumb for pushing it, but the curves are what counts and throughout the range of performance class wattages (30-40 here I picked) Qualcomm looks damn good in those ranges and better than AMD actually by 20-25%.

As for Apple vs Intel

Notice that the one M3 result is 50% more performant iso-power than any Lunar Lake at 21W (600 vs 400), or matches the MT performance of Lunar Lake around 40-45W (600 ish) at 1/2 the power. These are parts on the same N3B node, nearly the same size (139 for Intel vs like 146mm2 for the M3) with a 4 P + 4 E Core design, the same SLC cache size, blah blah. Intel also still has more total CPU area devoted to it than the M3 does, and actually more total cache for the P cores.

And it gets just blown out at 20W either way you slice it. Cinebench is FP but integer performance would follow a similar trend here.

AMD Entries:

Ryzen AI 9 365 (Yoga Pro 7 14ASP G9, 15W)

• Score: 589
• Wattage: 25.40W
• Performance/Watt: 23.2

Ryzen AI 9 365 (Yoga Pro 7 14ASP G9, 28W)

• Score: 787
• Wattage: 43.80W
• Performance/Watt: 18.0

Ryzen AI 9 HX 370 (Zenbook S16, 20W)

• Score: 767
• Wattage: 35.80W
• Performance/Watt: 21.4

Ryzen AI 9 365 (Yoga Pro 7 14ASP G9, 20W)

• Score: 688
• Wattage: 31.90W
• Performance/Watt: 21.4

Ryzen AI 9 HX 370 (Zenbook S16, 15W)

• Score: 672
• Wattage: 26.70W
• Performance/Watt: 25.2

Ryzen 7 8845HS (VIA 14 Pro, Quiet 20W)

• Score: 567
• Wattage: 27.70W
• Performance/Watt: 20.5

Intel Entries (SKUs ending in “V”):

Core Ultra 7 258V (Zenbook S 14 UX5406, Whisper Mode)

• Score: 406
• Wattage: 21.04W
• Performance/Watt: 19.3

Core Ultra 9 288V (Zenbook S 14 UX5406, Fullspeed Mode)

• Score: 598
• Wattage: 42.71W
• Performance/Watt: 14.0

Core Ultra 7 258V (Zenbook S 14 UX5406, Fullspeed Mode)

• Score: 602
• Wattage: 45.26W
• Performance/Watt: 13.3

Qualcomm Entries:

Snapdragon X Elite X1E-80-100 (Surface Laptop 7)

• Score: 897
• Wattage: 40.41W
• Performance/Watt: 22.2

Snapdragon X Elite X1E-78-100 (Vivobook S 15 OLED Snapdragon, Whisper Mode 20W)

• Score: 786
• Wattage: 36.10W
• Performance/Watt: 21.8

Snapdragon X Elite X1E-84-100 (Galaxy Book4 Edge 16)

• Score: 866
• Wattage: 39.10W
• Performance/Watt: 22.1

Apple Entry:

Apple M3 (MacBook Air 13 M3 8C GPU)

• Score: 601
• Wattage: 21.20W
• Performance/Watt: 28.3
 
I have heard that Lunar Lake in CB24 is broken so take those results with a grain of salt, but the QC vs AMD ones are good.
 
Well, conflicting rumors. You may be right now though.



Some report it’s still 2026 H1 and with Oryon V2, but that IP doens’t make sense. Oryon V2 is already shipping today (well very soon) at mass production. If it launched at the same time phones did or after, it would be a generation behind.

My guess is that it’s either early 2025 or summer 2025 and V2, or it’s V3 and late 2025/early 2026.
Okay maybe I'm stretching things but Qualcomm fiscal quarters start the previous year, so maybe this leak they mean H1 2026 could be September/October 2025. Maybe? Seems unlikely since they'd probably say Q1 rather than H1, but I dunno. Even September still seems late given the cores are available now and there was a report Glymur was already being tested a few weeks ago. But maybe they want to give themselves more time to get the software/drivers/accompanying SOC hardware right? Everyone thought highly of the CPU cores themselves in the last generation, a little late though they were, but it's everything else that needs the biggest boost. So maybe they're giving themselves more time to get that stuff right? Also it's possible if/when Nvidia-Mediatek releases WoA things will improve faster on the software compatibility front.
I had to gloat, but :).

FWIW, literally have an iPhone 16 Pro. Apple is still ahead, and the A18 and A18 Pro are still more impressive engineering, but it’s closer than ever before.


The main reason again I am excited is I don’t think Arm can get it done with their designs to the extent Apple can and expect that momentum will continue on Qualcomm’s end for Windows and for phones.

They have area efficiency (Apple does too actually), scalability, performance per watt and in the next generation I think they’ll get a lot closer on IPC, and with their clocking and efficiency much like Apple, I expect that means huge ST for Windows laptops without blowing power or area up as much as AMD and Intel, and next time they’ll have real E Cores.

I very much agree the addition of M-cores should be a big big boost - and the possibility of even S-cores if they have those too (although M-core die area seems very good already so I don't think they'll actually need S-cores, it's just the nomenclature is suggestive of those). Also glad to see two different SOCs being developed. They needed that as well. The original Elite was trying to cover too much ground with one SOC and no small cores. Doesn't seem to be any word on the GPU sizes for the two dies? But a 6+6 CPU could be a fanless M4 competitor - if not then maybe they'll have a binned version of it, say a 4+6 ;), that is appropriate for fanless designs.
The X Elite goes toe-to-toe with Lunar Lake on multiple fronts despite multiple disadvantages and a rushed product, and lower cost. AMD and Intel have nothing that can slot into $700-900 laptops with great ST and great battery life, but QC does, and it’s all about architecture.

Think about the next one lol.



People are underrating WoA vs AMD and Intel still for those reasons I think
Qualcomm (and Microsoft!) have got to execute better though - no more canceled dev kits, more native apps, better Prism, better dev support, better drivers. I do agree though that the Qualcomm CPU designers are doing their part.

I think people here were underestimating how competent these guys are just because Apple is still good. Human capital is huge in chip design. Sure, Qualcomm can act as a brick wall against that, but by all evidence they are not and did not want to, whereas with Intel internal fights and catabolism is just Tuesday.

But to some degree as I pointed out, their obstacles were one-time, the lawsuit, or getting in the groove of things with the X Elite which is probably why power wasn’t as good as we expected and there are multiple rumors and speculation on that. The fact that the turnaround was so significant and catapulted them ahead of the X925 for now is big.

That lawsuit is getting worse though. I very much doubt ARM and Qualcomm will go full nuclear, but that could still be a dampener going forwards.
It’s a real ball game.

Things do take time, so holding your breath from 2021 wasn’t wise and Cliff’s skepticism was warranted for sure, and the initial product was disappointing for the time (just less so relative to the cost and scalability imo).

But from the point we saw they were shipping product with a reasonable advantage on Intel/AMD and then putting them in *phones* right afterwards — that was a huge tell to me. They couldn’t put something in that was objectively worse or at least not a little better and we knew where the X925 and such were going.
 
GWIII also expressly hinted that they have another IPC upgrade coming which they really didn’t get with this to a big extent fwiw. I could be wrong though and it’s this:

V3 has the big upgrade for phones next fall

But at the same time Oryon V2 ships in laptops in the next chips next fall.

I could see it. It might explain why 4.8-5GHz is rumored. it’d still be great, just not ideal. A even a 10-20% IPC upgrade while keeping frequency constant for N3P and not increasing power too much would be killer for Windows laptops.
Cristiano Amon hinted during this year's Snapdragon Summit Day 1 Keynote that they have some big things to be announced at next year's Snapdragon Summit. My guess;

Snapdragon 8 Elite Gen 2
Snapdragon X Elite Gen 2

Both with next generation Oryon G3 cores (Pegasus-L and Pegasus-M), and crafted on TSMC N3P process.

______

Small nitpick: I think it's best we call it Oryon G3 (Oryon Gen 3), not Oryon V3 (Oryon Version 3).

The 'V' naming scheme came from that leaked Dell 2023 slide, and we now know that's clearly wrong.

Qualcomm is officially calling them Oryon 1st Gen, Oryon 2nd Gen, etc...
 
Okay maybe I'm stretching things but Qualcomm fiscal quarters start the previous year, so maybe this leak they mean H1 2026 could be September/October 2025. Maybe? Seems unlikely since they'd probably say Q1 rather than H1, but I dunno. Even September still seems late given the cores are available now and there was a report Glymur was already being tested a few weeks ago. But maybe they want to give themselves more time to get the software/drivers/accompanying SOC hardware right? Everyone thought highly of the CPU cores themselves in the last generation, a little late though they were, but it's everything else that needs the biggest boost. So maybe they're giving themselves more time to get that stuff right? Also it's possible if/when Nvidia-Mediatek releases WoA things will improve faster on the software compatibility front.
Yeah I just have a hard time seeing it be V2 in September next year. But here’s the thing. Even if it is, it will still be a very, very high performance and good laptop and they have frequency/power headroom.

But I really expect V3. If Glymur is being tested right now then based off some kind of 8-12 month range for their generation time then V3 makes total sense
I very much agree the addition of M-cores should be a big big boost - and the possibility of even S-cores if they have those too (although M-core die area seems very good already so I don't think they'll actually need S-cores, it's just the nomenclature is suggestive of those). Also glad to see two different SOCs being developed. They needed that as well. The original Elite was trying to cover too much ground with one SOC and no small cores. Doesn't seem to be any word on the GPU sizes for the two dies? But a 6+6 CPU could be a fanless M4 competitor - if not then maybe they'll have a binned version of it, say a 4+6 ;), that is appropriate for fanless designs.

it will be 6+6 and 12+6 by all reports yeah. And the Graphics will be good this time, with bigger sizing and a bigger bus for the 12+6 model.

The Oryon first generation without small cores was just really impressive though RE battery life. Think about what QC is going to do to Intel with an arch upgrade on sheer performance and even more battery life, probably still with reasonably competitive area and cost. I actually think we will see a breaking point and the gaps will be worse than now where they are similar thanks to LNL being more costly and big etc.
Qualcomm (and Microsoft!) have got to execute better though - no more canceled dev kits, more native apps, better Prism, better dev support, better drivers. I do agree though that the Qualcomm CPU designers are doing their part.

Agree but I mean realistically Nvidia will be the thing that pushes that through along with QC getting the GPU hardware and drivers to C+ or B-, right now I’d give it a fail in X Elite.


That lawsuit is getting worse though. I very much doubt ARM and Qualcomm will go full nuclear, but that could still be a dampener going forwards.
Agreed, but I mean it seems like it’s just negotiation. There’s not much to dampen — they’ll do bit higher royalties and still save money a one time payment or the whole show is over and QC calls it quits. Kind of an all in all out thing imo
 
I also have a little essay from Reddit a few of you might like on area and Lunar Lake vs the M2 (not even M3, it’s more M2 comparable in some ways), and QC vs Strix Point.

Also some CB24 MT data from the wall with the display off, just showing how solid the area and wattage efficiency can be here.
These look very similar to my own results and conclusions from similar analyses:
CB R24
Die shot analyses 1 and 2

Though as @mr_roboto pointed out, I got fooled by Intel's renaming of L1d to L0 and yet keeping a seemingly larger L1d which is more like L1.5d around.
I have heard that Lunar Lake in CB24 is broken so take those results with a grain of salt, but the QC vs AMD ones are good.
Really? On Linux or Windows too?
Yeah I just have a hard time seeing it be V2 in September next year. But here’s the thing. Even if it is, it will still be a very, very high performance and good laptop and they have frequency/power headroom.

But I really expect V3. If Glymur is being tested right now then based off some kind of 8-12 month range for their generation time then V3 makes total sense


it will be 6+6 and 12+6 by all reports yeah. And the Graphics will be good this time, with bigger sizing and a bigger bus for the 12+6 model.

The Oryon first generation without small cores was just really impressive though RE battery life. Think about what QC is going to do to Intel with an arch upgrade on sheer performance and even more battery life, probably still with reasonably competitive area and cost. I actually think we will see a breaking point and the gaps will be worse than now where they are similar thanks to LNL being more costly and big etc.


Agree but I mean realistically Nvidia will be the thing that pushes that through along with QC getting the GPU hardware and drivers to C+ or B-, right now I’d give it a fail in X Elite.



Agreed, but I mean it seems like it’s just negotiation. There’s not much to dampen — they’ll do bit higher royalties and still save money a one time payment or the whole show is over and QC calls it quits. Kind of an all in all out thing imo
My worry is that Qualcomm's partners will be reticent to release designs until the negotiations are done, potentially pushing back the release. Their partners won't want to release designs that could potentially get yanked even if that possibility is small.

Cristiano Amon hinted during this year's Snapdragon Summit Day 1 Keynote that they have some big things to be announced at next year's Snapdragon Summit. My guess;

Snapdragon 8 Elite Gen 2
Snapdragon X Elite Gen 2

Both with next generation Oryon G3 cores (Pegasus-L and Pegasus-M), and crafted on TSMC N3P process.

______

Small nitpick: I think it's best we call it Oryon G3 (Oryon Gen 3), not Oryon V3 (Oryon Version 3).

The 'V' naming scheme came from that leaked Dell 2023 slide, and we now know that's clearly wrong.

Qualcomm is officially calling them Oryon 1st Gen, Oryon 2nd Gen, etc...
I suppose if they are waiting for Gen 3 cores to release that could explain it, but there was a report that Glymur at least was already being tested. That could be wrong of course. It's just seems like 2026 is an awful long time to wait.
 
These look very similar to my own results and conclusions from similar analyses:
CB R24
Die shot analyses 1 and 2

Though as @mr_roboto pointed out, I got fooled by Intel's renaming of L1d to L0 and yet keeping a seemingly larger L1d which is more like L1.5d around.

Really? On Linux or Windows too?

My worry is that Qualcomm's partners will be reticent to release designs until the negotiations are done, potentially pushing back the release. Their partners won't want to release designs that could potentially get yanked even if that possibility is small.


I suppose if they are waiting for Gen 3 cores to release that could explain it, but there was a report that Glymur at least was already being tested. That could be wrong of course. It's just seems like 2026 is an awful long time to wait.
Yeah, but the 8 Elite has been tested for like 9 months now in leaks, seems plausible to me they’re already on G3, but who knows.
 
Yeah, but the 8 Elite has been tested for like 9 months now in leaks, seems plausible to me they’re already on G3, but who knows.
Development timeliness for smartphone SoCs are shorter than that for PC SoCs.

IIRC, silicon needs to be sampling to laptop OEMs atleast 12 months before it hits the market. Whereas, for smartphones it's about 6 months.

There's rumours that Intel is already testing Panther Lake, and that chip is slated for a 2025Q4 release.

I expect Snapdragon X Elite Gen 2 will be announced in 2025 Snapdragon Summit (October 2025), with laptops hitting the market in early 2026H1.
 
Weird story. I wonder what is going on here? It will be interesting to see what happens with other phones as retail units come out.

1730399022206.png

1730399046933.png
 
Weird story. I wonder what is going on here? It will be interesting to see what happens with other phones as retail units come out.

View attachment 32388
View attachment 32389
It’s probably Xiaomi-specific. QC doesn’t do this sort of thing in software, but OEMs will and Android lets them do more than they otherwise could
 
Qualcomm CEO Cristiano Amon dropped this teaser:



Timestamp = 5:40


Yep. This is probably the IPC upgrade GWIII mentioned. Frankly, it’s incredible the turnaround they had to 8 Elite.

Also I suspect this means the laptops will have the IPC upgrade & V3 tbh. Would be surprised if they don’t. Not for launching in 2025 Winter to really H1 2026. The difference will be that 3-9 month delay whereas more smartphones will launch immediately afterwards in October.
 
God the copium about this is so funny because it’s going to get worse. I kept saying that, while it isn’t certain obviously and companies can BS — those phone numbers — 4-4.4GHz casually without big IPC gains suggested huge power changes and yeah pretty much even more than I expected. When you look at GB6 curves, 8 Elite is doing 2500 at 3W or 3000 at 5.5W.

As for SpecInt (better indicator to some degree, FP is irrelevant largely), just like the X925 it hit an ~ 8 at 4W which is really the peak of the curve tradeoff. The 8.8-9 at 6-7.5W is insane and not worth it.

For reference Intel’s Lunar Lake does about an 8 in Geekerwan’s tests in SpecInt but at ~ 13-15W ish (again platform.)

Removing on-package from the phone and widening the bus would push the numbers up slightly, but not that much here, and we know Intel ain’t doing On-Package anymore with PTL so lol.

Suspect something about the fabric was a rush on the X Elite, power delivery is what I’ve heard. Arch and cache changes don’t at all save power like that they made (L1 reduction just allows higher clocks without blowing up area or other tweaks to it I think).
 
That said, while I think Qualcomm is really going to give us Store Brand (and that’s kind of too harsh at this pace but IMO no one else has) Apple SoCs next year and on —

I think the most impressive chip this year is the A18 & the A18 Pro, but specifically the A18.

Why?

Apple’s area actually isn’t that ridiculous at all if you include AMD/Intel whole cluster area instead of cherry picking private L2 for one core without L3, which is stupid due to their setups which inherently differ.

But even so, A18’s 8MB of L2 — same as the A14 — and 12MB of SLC — less than the A14’s 16MB — showed us just how far they can stretch it and be just fine for general ST and a phone. The results probably wouldn’t even be statistically significant across enough Spec & GB tests in terms of power.

Even the M2/3/4 show this basically, 16MB of L2 among 4 P cores, 4MB for E cores and an 8MB SLC. Pretty consistent. That’s…. Not that much.




Only thing I’d say is I suspect if you got some variant programs going back and forth in memory where the cache was occupied actively on the A18 Pro vs A18, you would probably be able to tease out some kind of hit to power/performance on an application if the other core was running hot with a larger working set (with 8MB vs 16MB L2 and 12 vs 24 MB of SLC.) Whether or not it’s enough to matter in a phone is a separate issue, and I suspect the answer is broadly no, and yes for some specific tasks (See media manipulation on the nicer iPhones) where Apple thinks the extra cache helps. Has to be something specific.
 
Also feel pretty vindicated saying Lunar Lake wasn’t impressive at all architecturally for the cost involved. The party looks so funny lol.

Intel can only pull that BS for as long as X86 software is a big deal for developers creators and gamers on Windows. It will erode thanks to MS, QC and soon & most importantly Nvidia next year, just taking time.


New ASIO low-latency audio drivers for Arm64 Windows is huge, relatedly.
 
Well, Qualcomm has to deliver a big upgrade with Oryon G3 to make X Elite G2 competitive.

The CPU (it's performance and efficiency) is really going to be their differentiating factor.

GPU isn't Qualcomm's strong point on PC. Intel/AMD/Nvidia are going to have better GPUs. (Oh yeah, there was a rumour that the Nvidia laptop SoC is coming in September 2025)

NPU could be anyone's ball. Pretty much impossible to guess how that will go next generation.

Qualcomm does have some strengths such as their mobile-derived ISP and their connectivity solutions (BT/WiFi/5G), but these will simply be cherries on top. The cake has to be made from the Oryon CPU.
 
Back
Top