Nintendo Switch Technical discussion [SOC = Tegra X1]

Except Nintendo has shown no interest in playing the graphics game and Nvidia has gotten out of developing mobile solutions so now any work for Nintendo would be just that, an SoC with a single customer and little hope of finding another.

This is a popular opinion, but I dont feel like its out of the question for Nvidia to create a custom processor for the next Switch. Nintendo paid big money to have IBM create a tri core PowerPC 750 processor and matched that to a custom AMD processor that used edram. The cost associated with developing custom processors gets pretty low when sales potential sits as high as 100 million units. I wouldn't be surprised to see the Switch successor use a processor that is a direct evolution of the X1. Switch 2 really only needs to get into the X1/PS4 realm of performance to offer a meaningful upgrade. Cut out the unnecessary hardware in the X1, and bump the GPU core count to 512. The A57 cores are actually better clock for clock than the Jaguar cores, so if hacked Switch units are anything to go by, bumping CPU core clock speeds to 1.75 Ghz is not a problem. Although is compatability isnt an issue, perhaps moving on to the A7* series CPU would be a much bigger improvement. Add in a 128bit memory bus with the newest LPDDR memory and now you have something that resembles a PS4 on the go. Portable consoles are always going to be a generation behind, but we have come so far that even a geration behind is impressive when its in the palm of your hands.
 
Nvidia stock has more than doubled YTD. What are they producing these days? Can't be a lot of growth in the PC 3D cards market right?

They had phone ambitions but never made it anywhere.

They are producing chipsets for cars but the world isn't exactly teeming with self-driving cars. Now they may be using Nvidia chips to control the dash display, like Audi does I believe. But that's not a fast-growing market either, car sales in general and high-end cars in particular.

So what are they so busy with that they couldn't knock off a new Switch SOC, though as successful as the Switch was, I'm guessing it didn't materially boost Nvidia's finances.

OK quick search, I found that Nvidia passed the market valuation of Intel when it's stock was $420, now it's $550.

For the first time, revenues from data center sales outpaced gaming hardware sales.

Meanwhile it has growing software sales tied to subscriptions:

The company is not yet ready to break out software sales and operating profits into separate line items in its quarterly reports. Instead, we have to look for clues like management comments in order to see the business category gaining momentum.

The software sales largely fall into three separate buckets.

  • AI: NVIDIA supports several software frameworks for artificial intelligence (AI) number-crunching, all optimized to take full advantage of GeForce graphics chips and Tesla (NASDAQ:TSLA) high-performance computing processors. The NVIDIA DIGITS neural network training system is an in-house solution for which NVIDIA offers enterprise-grade support services.
  • Self-driving cars: The NVIDIA DRIVE AGX Orin platform is a self-driving car system that's both built around NVIDIA's high-performance math processors and designed like a modular software product. Upgrades, bug fixes, and new features can be added through over-the-air software updates, and many features are available under subscription-style licensing terms. The Mercedes-Benz division of German automaker Daimler AG (OTC:DMLR.Y) (OTC:DDAI.F) recently committed to a DRIVE AGX partnershipthat will put NVIDIA-powered Mercedes cars on the road by 2024.
  • Data center networking: Through the recent acquisitions of Mellanox and Cumulus Networks, NVIDIA has become a significant force in the market for software-defined data center networks. This ultra-flexible network management approach comes with multiyear license and support agreements, pushing NVIDIA even further into the realm of software-based business.

OK, I can see the association with Tesla, whose own stock is going crazy, probably uplifts Nvidia.

Self-driving cars are mostly hype, nobody expects level 4 or 5 for at least a decade.

Data center, I can see, with the hype of the cloud.

Nvidia charges for AI frameworks which are meant to spur sales of its silicon? Nice business model.


Maybe Nintendo will have to pay a lot to get Nvidia to dedicate resources to making a bespoke chip. With high-flying stock, Nvidia shouldn't have problems attracting and retaining talent but console gaming may be far down their list of priorities.
 
Last edited:
Nvidia has said that their partnership with Nintendo was going to last for decades, so while its likely this was just PR fluff, but its also possible that there is certain agreements already set in place for future processors. Creating a "custom" processor doesnt mean its comprised of bespoke components. Piecing together a SOC comprised of a very mature architecture matched to licenced ARM CPU cores just doesn't scream high priced. Backwards compatibility will also be a big factor. If Nintendo wants that to be a feature, sticking with Nvidia is a guarantee. This continued partnership will mean Nintendo will likely end up with a processor less capable than they could have gotten for the same money from a mobile SOC manufacture, but its Nintendo, they value a lot of things above peak performance per dollar.
 
It is but those are high power parts (30-60 W) iirc, not suitable at all for a mobile platform, I mean the launch Switch battery life was marginal for a mobile device as it was. Someone posted their most recent platform roadmap here recently that included Tegra, Xavier, and Odin but I cannot for the life of me find it now. It showed a new Tegra class platform for 2022ish iirc which seems like the best fit for Nintendo's needs as more or less everything else Nvidia is doing in the ARM SoC space is AI and automation focused (60W!) and a poor fit for a handheld device.
 
Nintendo paid big money to have IBM create a tri core PowerPC 750 processor and matched that to a custom AMD processor that used edram.
Nintendo paid to have IBM make them a 3-core CPU based on a core IP that was turning 14 years old by the time the Wii U went to market, and AMD for a GPU based on the old VLIW5 DX10 architecture.
"Big money" is probably not the word to use here. Nintendo hasn't used state-of-the-art aechitectures since the gamecube.


The A57 cores are actually better clock for clock than the Jaguar cores
Are there public benchmarks that corroborate this?
 
Are there public benchmarks that corroborate this?
I see this a lot and never found anything that can back this up, a big part of the problem is most Jaguar products were out of the market by the time A57 was on the market so there's not any direct comparisons between both. Best x86 vs ARM comparison article I've found was this Jaguar vs A15 from notebookcheck.net article from 2013 but to be honest without a reliable way to try and eliminate variables such as varying Android builds/skins etc I think these comparisons are of very limited usefulness. ARM was in the final stages of completely eliminating x86 in low cost, low power devices where Atom and Jaguar used to dominate even in 2013 so by the time A57 rolled around it just never had to compete versus x86.
 
Nintendo paid to have IBM make them a 3-core CPU based on a core IP that was turning 14 years old by the time the Wii U went to market, and AMD for a GPU based on the old VLIW5 DX10 architecture.
"Big money" is probably not the word to use here. Nintendo hasn't used state-of-the-art aechitectures since the gamecube.



Are there public benchmarks that corroborate this?

Im just saying that Nintendo paid a good chunk of change to have a PPC750 customized into a tri core CPU, something that IBM never did with that series of processor. It was truly a one off custom piece. That may have been a poor investment, but it just goes to show that Nintendo makes decisions that arent made with the intent of getting the best performance or lowest cost.

Tom at DF is the one who talked about the A57 CPU cores having better IPC than the Jaguar cores. I believe it was during the Witcher 3 tech review.
 
Im just saying that Nintendo paid a good chunk of change to have a PPC750 customized into a tri core CPU, something that IBM never did with that series of processor. It was truly a one off custom piece. That may have been a poor investment, but it just goes to show that Nintendo makes decisions that arent made with the intent of getting the best performance or lowest cost.

Tom at DF is the one who talked about the A57 CPU cores having better IPC than the Jaguar cores. I believe it was during the Witcher 3 tech review.
Nintendo also paid for a custom SoC to be developed for the 3DS (and another one for the "New 3DS"), using IP that was unique enough to ensure there were no off-the-shelf parts for it.
 
Is there any Nvidia GPU which can have 2~3x performance "per watt" if
Switch 4k model is real?
No not in their road map today for the mobile space. Compared to the current architecture in the X1 the GPU is pretty old but they would need a lot more in the CPU side also. NV has been investing in non-mobile ARM parts for a while (their AI and car stuff).

Handheld 4K = no.
Docked 4K also no
4K upscale from 1080p with native UI elements seems pretty likely.
Not discounting any title being "native 4k" but I feel pretty confident none of the 3D fare like Zelda will get native 4k
 
Uh any idea how nintendo manages to make switch's SBC Bluetooth almost lag free?

I can't detect its lag at all. Despite I can easily detect it (albeit minimal, seems around 100ms) on windows pc with Intel BT, xiaomi android phone (only for gaming, as videos automatically got delay adjusted to sync with the audio), and lg CX oled TV (only in testing mode, as the TV automatically compensate for audio lag for normal use) .
 
Back
Top