• 0 Posts
  • 16 Comments
Joined 1 year ago
cake
Cake day: October 25th, 2023

help-circle
  • I mean it just seems sort of insane to use 3nm silicon companies like Nvidia are desperate to sell GPUs costing tens of thousands with margins of well over 500% on… selling a smart TV box? Saving a few dimes on bandwidth for the minority of users that are Apple TV+ subscribers? Helping push games which most Apple TV users don’t play?

    I wouldn’t be shocked if the last gen chip can just brute force av1 decode in software too. It just seems hilariously uneconomical.







  • 36gb of vram on the 384-bit bus would be fantastic, yet I’m somehow sceptical when Nvidia sells the 48gb A6000 for a $6800 MSRP. Even without benefits like nvlink, a 36gb card ought to cannibalise Nvidia’s productivity cards quite a lot. I don’t think Nvidia would actually be TOTALLY opposed to this if they could produce enough 5090’s to not sell out of them since it would help entrench Nvidia’s CUDA moat, but I don’t think Nvidia is going to be capable of pulling that off.

    It’s not impossible we see 36gb 5090 and 72gb a7000 or whatever. I’m just not holding my breath especially when AMD doesn’t seem to have much in the pipeline to even compete with a 24gb model.



  • They can technically make the emulator and have. It is hard to think of a company more qualified to do so than Microsoft, they’re frankly more equipped than Apple is.

    The problem Microsoft has more broadly is Apple is a company which has set the expectation that they don’t do legacy support. Apple is a company which has set the expectation that they will change things and their customers will pay the cost. So they can just straight up say “in 2 years, we won’t sell computers which use x86 anymore, transition now” and everybody does it and they only see higher sales.

    Microsoft is a company which people use because they have outstanding legacy support and save their customers money through supporting 10 year old line of business applications at their expense. If they move off x86 in the same way Apple did, they will bleed customers to Linux/ChromeOS/MacOS/Android/iPadOS etc. etc. So they’re essentially forced to support ARM and x86 concurrently. That results in every developer going “Well, more people are using x86, and a lot less people are using ARM, so I’ll just develop for x86 only and ARM users can emulate”. This results in the ARM experience being shit but there’s nothing Microsoft can do about it even though not transitioning more forcibly will kill Windows market share in the long term. It’s just not worth it to force things especially since Windows is doomed to die in slow motion regardless.


  • Pushing hard with ROCm?

    There are millions of Devs who develop for CUDA. Nvidia I believe has north of a thousand (can’t remember if it’s like 1 or 2 thousand) people working on Cuda. CUDA is 17 years old. There is SO MUCH work already done in CUDA, Nvidia is legit SO far ahead and I think people really underestimate this.

    If AMD hired like 2000 engineers to work on ROCm they would still take maybe 5 years to get to where Nvidia is now, and still be 5 years behind Nvidia. Let’s not even get into the magnitudes more CUDA GPUs floating around out there compared to ROCm GPUs, because CUDA GPUs started being made earlier at higher volumes and even really old stuff is still usable for learning/home lab. As far as I know, they’re hiring a lot less, they just open sourced it and are hoping they can convince enough other companies to write stuff for ROCm.

    I don’t mean to diminish AMD’s efforts here, Nvidia is certainly scared of ROCm, ROCm I expect to make strides in the consumer market in particular as hobbyists try and get their cheaper AMD chips to work with Diffusion models and whatever. When it comes to more enterprise facing stuff though CUDA is very very far ahead and the lead is WIDENING and the only real threat to that status quo is that there literally are not enough NVIDIA GPUs to go around.


  • Nvidia is all hands on deck going pedal to the metal right now trying to stop AMD from gaining any market share right now. They’re worried about becoming a victim of their own success here and the AI boom allowing AMD to gain a foothold in AI with their open source strategy. They’re also worried about Intel and Google for similar reasons.

    Nvidia is quite the formidable foe especially compared to Intel, and they have a massive headstart because they have a LOT of advantages beyond having merely the best hardware on the market, but I’m still a bit bullish on AMD’s chances here.




  • The 4080 + i9-13900hx laptop has more total memory, more total bandwidth, and more total transistors, with the only real downside being that system resources are split across two processors on a bandwidth constrained bus. If correctly optimised, there are almost no multi-threaded workloads where I would expect the m3 Pro to actually outperform a 4080 + i9-13900hx in terms of performance itself.

    I don’t even nessecarily see the m3 pro as being a serious workstation processor. I see it as being almost a thin and light processor for people with moderate workloads like non-fanciful 4k editing or normalish development work. The battery life you’re going to get from it compared to a system running two independent processors on less recent lithography is going to be a LOT better. The single threaded performance is a lot better. The performance is still good enough for a heavy task here and there.

    I get the compulsion to compare roughly price equivalent laptops to eachother, but I don’t even feel like these two machines aren’t even really optimised for the same use case, so comparing their workstation performance since the computers are about the same price and being like “looks like Apple still has catching up to do” is almost missing the point. I feel like it makes more sense to compare this to the Max, even though the price isn’t the same. I’d compare this laptop to a cheaper high end AMD APU laptop because those laptops are a lot closer to eachother philosophically.