by dvcoolarun on 2/23/24, 1:52 PM with 106 comments
by Yenrabbit on 2/24/24, 6:21 PM
For those talking about breakeven points and cheap cloud compute, you need to factor in the mental difference it makes running a test locally (which feels free) vs setting up a server and knowing you're paying per hour it's running. Even if the cost is low, I do different kinds of experiments knowing I'm not 'wasting money' every minute the GPU sits idle. Once something is working, then sure scaling up on cheap cloud compute makes sense. But it's really, really nice having local compute to get to that state.
by abra0 on 2/24/24, 2:42 PM
by infogulch on 2/24/24, 1:51 PM
by bick_nyers on 2/24/24, 8:02 PM
I have a 3U supermicro server chassis that I put an AM4 motherboard into, but I'm looking at upgrading the Mobo so that I can run ~6 3090s in it. I don't have enough physical PCIE slots/brackets in the chassis (7 expansion slots), so I either need to try to do some complicated liquid cooling setup to make the cards single slot (I don't want to do this), or I need to get a bunch of riser cables and mount the GPU above the chassis. Is there like a JBOD equivalent enclosure for PCIE cards? I don't really think I can run the risers out the back of the case, so I'll likely need to take off/modify the top panel somehow. What I'm picturing in my head is basically a 3U to 6U case conversion, but I'm trying to minimize cost (let's say $200 for the chassis/mount component) as well as not have to cut metal.
by kaycebasques on 2/24/24, 3:24 PM
Am also inspired by embedded developers for the same reason
by neilv on 2/24/24, 5:10 PM
by Uehreka on 2/24/24, 5:00 PM
I would prefer a tutorial on how to do this.
by whoisthemachine on 2/24/24, 4:52 PM
My box has a Gigabyte B450M, Ryzen 2700X, 32GB RAM, Radeon 6700XT (for gaming/streaming to steam link on Linux), and an "old" Geforce GTX 1650 with a paltry 6GB of RAM for running models on. Currently it works nicely with smaller models on ollama :) and it's been fun to get it set up. Obviously, now that the software is running I could easily swap in a more modern NVidia card with little hassle!
I've also been eyeing the b450 steel legend as a more capable board for expansion than the Gigabyte board, this article gives me some confidence that it is a solid board.
by smokeydoe on 2/24/24, 5:45 PM
by 0x20cowboy on 2/24/24, 8:01 PM
The main benefit is you can shut off nodes entirely when not using them, and then when you turn them back on they just rejoin the cluster.
It also helps managing different types of devices and workloads (tpu vs gpu vs cpu)
by akasakahakada on 2/25/24, 1:55 AM
2 x RTX4090 workstation guide
You can put two aircooled 4090 in the same ATX case if you do enough research.
https://github.com/eul94458/Memo/blob/main/dual_rtx4090works...
by cyanydeez on 2/24/24, 2:48 PM
those rigs need pcie riser slots that are also limited.
looks like the primary value is the rig and the cards. they'll need another 1-2k for a thread ripper and then the riser slots.
by nirav72 on 2/24/24, 4:52 PM
by whimsicalism on 2/24/24, 5:25 PM
inb4 there are no cloud 3090s: yes there are, just not in formal datacenters
by jeffybefffy519 on 2/24/24, 8:33 PM
by gigatexal on 2/24/24, 5:03 PM