by shubham_saboo on 8/24/23, 5:53 AM with 28 comments
by alsodumb on 8/24/23, 7:00 AM
by weinzierl on 8/24/23, 12:52 PM
32k context length sounds nice of course, and it seems to be common to call the just fine-tuned models like that. I think it is more of a marketing thing and we really should distinguish between the context length of the pre-trained model and the fine-tuned model, with the latter being the default meaning of context length.
by supermatt on 8/24/23, 7:03 AM
Edit: No mention of it being open source in the linked article. Maybe the title here is just wrong? @dang
by vekker on 8/24/23, 9:13 AM
by Havoc on 8/24/23, 9:47 AM
i.e. all other things being equal is a 8k model better at math than a 32k model