by blackcat201 on 5/8/21, 5:38 PM with 37 comments
by throwawaybbq1 on 5/9/21, 12:10 AM
For the non-DL crowd, Transformers are a tsunami in deep learning for the past few years. They are topping benchmarks in many subfields. I do research professionally and this work is amazingly useful for people like me.
by fao_ on 5/8/21, 9:57 PM
https://github.com/MrMEEE/ironhide
Optimus Support for Linux Through VirtualGL - PPA version also available
That's... great. So it's doing something with GL, and it's running on Linux, but uhhh. my branch of the original bumblebee project..
What is Optimus? What is Bumblebee? The trick of it is that it links to a blog where neither of these terms are ever explained. Maybe it's to just look impressive on someone's CV? How could I even tell the difference?Likewise for this project, all you need in the README is one line that's like:
X-Transformers is a re-implementation of Machine Learning Transformers that has been built based on experimental Arxiv papers
It's a one-line fix but it'll stop people like me being confused as to whether or not you're implementing a new HTTP headerby bratao on 5/9/21, 2:24 AM
About X-Transformers, it is a very great piece of engineering that implemented almost of all possible improvements in transformers. But according to my experience and Phil himself, only the Feedforward GLU and RoPe (Rotary Positional Embeddings) works (or to be fair, they show improvements in more general use-cases)
by argvargc on 5/8/21, 9:19 PM
by adontz on 5/8/21, 10:25 PM
by bravura on 5/9/21, 9:57 AM
by krick on 5/8/21, 10:43 PM
by mrfusion on 5/8/21, 8:48 PM
by shayankh on 5/8/21, 9:44 PM