by alexei_rudak on 8/1/23, 6:53 PM with 2 comments
I tried to make several small language models using 3-5 million of parallel sentence dataset to make specific linguistic transformation and see much potential there. The difference in performance between ChatGPT4 and my small models is almost 1000 times.
Anyone do such things ?
by omeze on 8/1/23, 7:08 PM
Fyi me and a friend made a demo app of tinystories we just submitted: https://news.ycombinator.com/item?id=36960333 - from testing it out, anything with <500ms latency and 15 tok/sec feels real-time, and that makes a world of difference in UX