by fbnbr on 5/7/24, 9:35 PM with 4 comments
by robertclaus on 5/8/24, 1:48 AM
An example with different fine-tuned models (especially smaller/cheaper ones) would probably be more interesting than running against a bunch of similar foundation models. For example throwing in some code-generation models and demonstrating that it picks those for coding problems.
by PeterStuer on 5/8/24, 3:05 PM
I was shocked how much better CodeQwen1.5 was at Python compared to ChatGPT 4.
by rfc on 5/7/24, 10:57 PM
Very cool... nice work.