by DreamFlasher on 3/15/23, 8:57 PM with 77 comments
by fpgaminer on 3/15/23, 10:49 PM
PyTorch 2.0 comes with a few different efficient transformer implementations built-in. And unlike 1.13, they work during training and don't require specific configurations. Seemed to work just fine during my pre-release testing. Also, having it built into PyTorch might mean more pressure to keep it optimized. As-is xformers targets A100 primarily, with other archs as an afterthought.
And, as promised, `torch.compile` worked out of the box, providing IIRC a nice ~20% speed up on a ViT without any other tuning.
I did have to do some dependency fiddling on the pre-release version. Been looking forward to the "stable" release before using it more extensively.
Anyone else seeing nice boosts from `torch.compile`?
by mardifoufs on 3/15/23, 10:22 PM
>Due to lack of Python 3.11 support for packages that PyTorch depends on, including NumPy, SciPy, SymPy, Pillow and others on the Anaconda platform. We will not be releasing Conda binaries compiled with Python 3.11 for PyTorch Release 2.0. The Pip packages with Python 3.11 support will be released, hence if you intend to use PyTorch 2.0 with Python 3.11 please use our Pip packages.
It really sucks that anaconda always lags behind. I know the reasoning*, and I know it makes sense for what a lot of teams use it for... but on our side we are now looking more and more into dropping it since we are more of an R&D team. We already use containers for most of our pipelines, so just using pip might be viable.
*Though I guess Anaconda chewed more than it can handle w.r.t managing an entire Python universe, and keeping up to date. Conda-forge is already almost a requirement but using the official package (with pip, in this case) has its own benefits for very complex packages like pytorch.
by brucethemoose2 on 3/15/23, 9:09 PM
Also, I have been using torch.compile for the Stable Diffusion unet/vae since February, to good effect. I'm guessing similar optimizations will pop up for LLaMA.
by singularity2001 on 3/16/23, 4:42 AM
That's (for me) the biggest reason why tensor flow fell out of flavor: the API broke too often (not just between tf 1 and 2)
by simonw on 3/15/23, 9:14 PM
by lucasap on 3/16/23, 8:15 AM
> Python 1.8 (deprecating Python 1.7)
> Deprecation of Cuda 11.6 and Python 1.7 support for PyTorch 2.0
It is clearly supposed to be python 3.8 and 3.7 respectively.
by tormeh on 3/16/23, 1:38 AM
by yumraj on 3/15/23, 11:44 PM
by marviel on 3/16/23, 1:13 PM
by mdaniel on 3/15/23, 9:03 PM