this post was submitted on 19 May 2025
56 points (100.0% liked)

LocalLLaMA

2972 readers
106 users here now

Welcome to LocalLLaMA! Here we discuss running and developing machine learning models at home. Lets explore cutting edge open source neural network technology together.

Get support from the community! Ask questions, share prompts, discuss benchmarks, get hyped at the latest and greatest model releases! Enjoy talking about our awesome hobby.

As ambassadors of the self-hosting machine learning community, we strive to support each other and share our enthusiasm in a positive constructive way.

founded 2 years ago
MODERATORS
 

"While the B60 is designed for powerful 'Project Battlematrix' AI workstations sold as full systems ranging from $5,000 to $10,000, it will carry a roughly $500 per-unit price tag."

top 18 comments
sorted by: hot top controversial new old
[–] possiblylinux127@lemmy.zip 1 points 4 hours ago
[–] billwashere@lemmy.world 6 points 8 hours ago (1 children)

No they didn’t. Or at least no one will ever actually be able to purchase one. I’m still waiting on an in stock notification for the last one.

[–] aow@sh.itjust.works 1 points 8 hours ago (1 children)

Not sure where you are in the world, but I got a B580 from b&h in February without issues. Just checked the site once a day when I knew I wanted to buy one.

[–] billwashere@lemmy.world 1 points 6 hours ago (1 children)

That was where I signed up for stock alerts. The only alert I got was about the ASRock B580 no longer being carried, at all. I wanted one of the actual Intel cards for somewhere close to the MSRP not a card for 2x MSRP plus a bundled power supply.

[–] aow@sh.itjust.works 1 points 6 hours ago* (last edited 6 hours ago)

Yeah I think their alert system sucks or something. Manual checks on that, microcenter, and best buy are probably your best bet for any new releases going forward.

Edit: I bought an Intel one, 250$ or whatever it was, been very happy with it

[–] brucethemoose@lemmy.world 1 points 8 hours ago (1 children)

With what 3090s are going for, this is good.

What GDDR ICs are they using? If blackwell can top out at 96GB, seems like the Arch could do at least 32GB?

[–] possiblylinux127@lemmy.zip 1 points 3 hours ago

I think it comes down to not having enough space on the board

[–] simple@lemm.ee 8 points 14 hours ago* (last edited 14 hours ago)

$299 with 16gb of memory

So 450-500 dollars once they instantly go out of stock and nobody can find them anywhere. I want to be excited but MSRP has been a joke since covid. Either way, the cards look good. i hope they can iron out the driver issues this time around.

[–] hendrik@palaver.p3x.de 8 points 15 hours ago* (last edited 8 hours ago) (1 children)

~~Nice.~~ I always first look at the memory bandwidth if it's about AI. ~~And seems with a 224bit bus, they've done a better job than previous cards you'd find in that price segment.~~

[–] brucethemoose@lemmy.world 2 points 8 hours ago (1 children)

224bit bus

Where do you see that? I thought the die was 192-bit at most.

There's is a specification for 224 GB/S.

[–] hendrik@palaver.p3x.de 3 points 8 hours ago* (last edited 8 hours ago) (1 children)

Thank you very much for the correct information. I googled it and took the number from some random pc news page. Either they got it wrong or I might need new glasses. Nonetheless, 128 or 192-bit is what Intel has on their website. I wish they'd do more for cards with 16GB of VRAM or more. I think two hundred and something GB/s is about what Nvidia, AMD and everyone already did in their previous generation of graphics cards.

[–] brucethemoose@lemmy.world 2 points 8 hours ago (1 children)

Prices for AMD/Nvidia (except maybe a used AMD 7900 XTX) are so awful that this is still a good deal, no matter how much bandwidth it has. For pure text LLM usage, capacity is king.

Intel's hands are tied buy what silicon they have available, unfortunately.

[–] hendrik@palaver.p3x.de 1 points 8 hours ago* (last edited 8 hours ago) (1 children)

Hmm. I could buy a (new) Radeon 7600 XT right now for around 330€... that should be only slightly more than $300 plus VAT, and that also has 16GB of VRAM and a similar (slightly faster?) memory interface?

[–] brucethemoose@lemmy.world 2 points 7 hours ago* (last edited 7 hours ago) (1 children)

The better choice depends on which software stack you intend to wrangle with, how long you intend to keep the cards, and your usage patterns, but the B580 is the sorta newer silicon.

Exllamav3 is the shit these days (as you can fully offload 32Bs in 16GB with very little loss), and it's theoretically getting AMD support way before Intel (unless Intel changes that).

...Also, 2x 3060s may be a better option, depending on price.

[–] hendrik@palaver.p3x.de 2 points 7 hours ago* (last edited 6 hours ago) (1 children)

It's difficult to make that decision. I mainly tinker with generative AI. Try a chatbot or agent or try creative writing with it, mess with FramePack or LTX-video, or explore some text to speech or whatever I find interesting when I got some time to spare.
Obviously I wouldn't spend a lot of money just to mess around. So yeah. I currently just rent cloud GPUs by the hour and I'm fine with that. Once we get a very affordable and nice AI card with lots if fast VRAM, I think I'm going to buy one. But I'm really not sure if this one or any previous generation Radeon is what I'm looking for. Or me spending quite some time on ebay to find some old 30x0. And as a Linux user I'm not really fond of the Nvidia drivers, so that also doesn't make it easier.

[–] brucethemoose@lemmy.world 2 points 5 hours ago* (last edited 5 hours ago)

Oh yeah, you will run into a ton of pain sampling random projects on AMD/Intel. Most "experiments" only work out of the box on Nvidia. Some can be fixed, some can't.

A used 3090 is like gold if you can find one, yeah.

And yes, I sympathize with Nvidia being a pain on linux... though it's not so bad if you just output from your IGP or another card.

And yes, stuff rented from vast.ai or whatever is cheap. So are APIs. TBH thats probably the way to go if budget is a big concern, and a 24GB B60 is out of the cards.

[–] jia_tan@lemmy.blahaj.zone 4 points 14 hours ago
[–] sith@lemmy.zip 2 points 13 hours ago

Too bad IPEX-LLM is dragging their feet.