by lolinder on 2/17/24, 5:44 PM
Mistral's process for releasing new models is
extremely low-information. After getting very confused by this link I tried looking for a link that has
any better information, and there just isn't one.
I thought Mixtral's release was weird when they just pasted a magnet link [0] into Twitter with no information, but at least people could download and analyze it so we got some reasonable third-party commentary in between that and the official announcement. With this one there's nothing at all to go on besides the name and the black box.
[0] https://news.ycombinator.com/item?id=38570537
by mattpavelle on 2/17/24, 12:25 PM
For those unfamiliar with the LMSys interface:
Click/tap on "Direct Chat" in the top tab navigation and you can select "mistral-next" as model.
by twobitshifter on 2/17/24, 2:11 PM
AIExplained on youtube has guessed that Gemini 1.5 pro is taking Mistral’s accurate long content retrieval and Google just scaled it as much as they could. The Gemini 1.5 pro paper has a citation back to the last mistral paper in 2024.
by mrfakename on 2/17/24, 5:10 PM
Note that it's actually "Mistral Next" not "Mixtral Next" - so it isn't necessarily a MoE. For example, an early version of Mistral Medium (Miqu) was not a MoE but instead a Llama 70B model. I wonder how many parameters this one has
by bloopernova on 2/17/24, 3:36 PM
Slightly related question: what's a good coding LLM to run on a 4070 12GB card?
Also, do coding LLMs use treesitter to "understand" code?
by aunetx on 2/17/24, 1:52 PM
It's quite funny to use! It is better when speaking French than chat gpt3.5 on my opinion
by justanotherjoe on 2/17/24, 2:35 PM
wow, this might be the best LLM that i've used in terms of phrasing and presenting the answers.
by AnujNayyar on 2/17/24, 11:16 AM
No indication that this a MoE (Mistral not Mixtral).
Very exciting nevertheless, here’s hoping the bless the OS community once again!
by ismailmaj on 2/17/24, 12:11 PM
Could it be Mistral Large?
This beats GPT-4 on my personal test.
by tmikaeld on 2/17/24, 12:18 PM
This was linked randomly on Mistrals Discord chat, nothing "official" yet.
It's a preview of their newest prototype model.
To use it, click "Direct Chat" tab and choose "Mistral next"
by ccwilson10 on 2/18/24, 5:55 AM
I used this but, upon asking which model it is, it replied as being a "fine-tuned version of GPT 3.5". Any clue why? In a second chat it replied "You're chatting with one of the fine-tuned versions of the OpenAssistant model!".
by vitorgrs on 2/17/24, 6:18 PM
From my tests, it did better than Gemini Ultra on a few reason/logic questions.
by apapapa on 2/17/24, 6:32 PM
The Together.AI logo at the bottom is very hard to read... (Dark gray on black)
by redder23 on 2/17/24, 6:01 PM
You can literally type "woke shit" in and you get woke shit out. I am so impressed.
by xeckr on 2/17/24, 6:50 PM
As someone who has only been using GPT-4 since its release, I am pleasantly surprised by how far open LLMs have come.