by andrewon on 4/28/23, 2:26 AM with 84 comments
by jiggawatts on 4/29/23, 10:19 PM
None of the Python dependencies are strongly versioned, and “something” happened to the CUDA compatibility of one of them about a month ago. The original developers “got lucky” but now nobody else can compile this stuff.
After years of using only C# and Rust, both of which have sane package managers with semantic versioning, lock files, reproducible builds, and even SHA checksums the Python package ecosystem looks ridiculously immature and even childish.
Seriously, can anyone here build a docker image for running these models on CUDA? I think right now it’s borderline impossible, but I’d be happy to be corrected…
by doodlesdev on 4/29/23, 10:48 PM
> Our system thinks you might be a robot!
We're really sorry about this, but it's getting harder and harder to tell the difference between humans and bots these days.
Yeah, fuck you too. Come on, really, why put this in front of a _blog post_? Is it that hard to keep up with the bot requests when serving a static page?by vessenes on 4/29/23, 10:35 PM
by simonw on 4/29/23, 8:15 PM
The same team that built that iPhone app - MLC - also got Vicuna running directly in a web browser using Web GPU: https://simonwillison.net/2023/Apr/16/web-llm/
by brucethemoose2 on 4/28/23, 3:34 AM
by brianjking on 4/30/23, 1:46 AM
Secondly, most of us can't even use the model for research or personal use, given the license.
by FloatArtifact on 4/29/23, 8:52 PM
by foobarbecue on 4/30/23, 12:44 AM