by luiscosio on 7/23/24, 2:47 PM with 269 comments
by dang on 7/23/24, 5:56 PM
Open source AI is the path forward - https://news.ycombinator.com/item?id=41046773 - July 2024 (278 comments)
by lelag on 7/23/24, 3:25 PM
Quick comparison with GPT-4o:
+----------------+-------+-------+
| Metric | GPT-4o| Llama |
| | | 3.1 |
| | | 405B |
+----------------+-------+-------+
| MMLU | 88.7 | 88.6 |
| GPQA | 53.6 | 51.1 |
| MATH | 76.6 | 73.8 |
| HumanEval | 90.2 | 89.0 |
| MGSM | 90.5 | 91.6 |
+----------------+-------+-------+
by zone411 on 7/23/24, 8:35 PM
GPT-4o 30.7
GPT-4 turbo (2024-04-09) 29.7
Llama 3.1 405B Instruct 29.5
Claude 3.5 Sonnet 27.9
Claude 3 Opus 27.3
Llama 3.1 70B Instruct 26.4
Gemini Pro 1.5 0514 22.3
Gemma 2 27B Instruct 21.2
Mistral Large 17.7
Gemma 2 9B Instruct 16.3
Qwen 2 Instruct 72B 15.6
Gemini 1.5 Flash 15.3
GPT-4o mini 14.3
Llama 3.1 8B Instruct 14.0
DeepSeek-V2 Chat 236B (0628) 13.4
Nemotron-4 340B 12.7
Mixtral-8x22B Instruct 12.2
Yi Large 12.1
Command R Plus 11.1
Mistral Small 9.3
Reka Core-20240501 9.1
GLM-4 9.0
Qwen 1.5 Chat 32B 8.7
Phi-3 Small 8k 8.4
DBRX 8.0
by foundval on 7/23/24, 3:36 PM
If you want to learn more, there is a writeup at https://wow.groq.com/now-available-on-groq-the-largest-and-m....
(disclaimer, I am a Groq employee)
by netsec_burn on 7/23/24, 3:09 PM
Statement from Mark: https://about.fb.com/news/2024/07/open-source-ai-is-the-path...
by meetpateltech on 7/23/24, 3:00 PM
https://about.fb.com/news/2024/07/open-source-ai-is-the-path...
by ajhai on 7/23/24, 6:12 PM
If you want a playground to test this model locally or want to quickly build some applications with it, you can try LLMStack (https://github.com/trypromptly/LLMStack). I wrote last week about how to configure and use Ollama with LLMStack at https://docs.trypromptly.com/guides/using-llama3-with-ollama.
Disclaimer: I'm the maintainer of LLMStack
by primaprashant on 7/23/24, 3:14 PM
by CGamesPlay on 7/24/24, 1:42 AM
Examples: OpenAI's GPT 4o-mini is second only to 4o on LMSys Overall, but is 6.7 points behind 4o on MMLU. It's "punching above its weight" in real-world contexts. The Gemma series (9B and 27B) are similar, both beating the mean in terms of ELO per MMLU point. Microsoft's Phi series are all below the mean, meaning they have strong MMLU scores but aren't preferred in real-world contexts.
Llama 3 8B previously did substantially better than the mean on LMSys Overall, so hopefully Llama 3.1 8B will be even better! The 70B variant was interestingly right on the mean. Hopefully the 430B variant won't fall below!
by kingsleyopara on 7/23/24, 3:38 PM
by Workaccount2 on 7/23/24, 4:30 PM
by AaronFriel on 7/23/24, 3:10 PM
Open source models are very exciting for self hosting, but the per-token hosted inference pricing hasn't been competitive with OpenAI and Anthropic, at least for a given tier of quality. (E.g.: Llama 3 70B costing between $1 and $10 per million tokens on various platforms, but Claude Sonnet 3.5 is $3 per million.)
by primaprashant on 7/23/24, 2:59 PM
[1]: https://github.com/meta-llama/llama-models/blob/main/models/...
[2]: https://github.com/meta-llama/llama-recipes/blob/main/recipe...
by dado3212 on 7/23/24, 4:35 PM
Have other major models explicitly communicated that they're trained on synthetic data?
by jcmp on 7/23/24, 3:06 PM
by anotherpaulg on 7/24/24, 7:15 AM
https://aider.chat/docs/leaderboards/
77.4% claude-3.5-sonnet
75.2% DeepSeek Coder V2 (whole)
72.9% gpt-4o
69.9% DeepSeek Chat V2 0628
68.4% claude-3-opus-20240229
67.7% gpt-4-0613
66.2% llama-3.1-405b-instruct (whole)
by sagz on 7/23/24, 3:41 PM
by ofou on 7/24/24, 4:12 AM
Llama 3 Training System
19.2 exaFLOPS
_____
/ \ Cluster 1 Cluster 2
/ \ 9.6 exaFLOPS 9.6 exaFLOPS
/ \ _______ _______
/ ___ \ / \ / \
,----' / \`. `-' 24000 `--' 24000 `----.
( _/ __) GPUs GPUs )
`---'( / ) 400+ TFLOPS 400+ TFLOPS ,'
\ ( / per GPU per GPU ,'
\ \/ ,'
\ \ TOTAL SYSTEM ,'
\ \ 19,200,000 TFLOPS ,'
\ \ 19.2 exaFLOPS ,'
\___\ ,'
`----------------'
by unraveller on 7/23/24, 4:27 PM
by sfblah on 7/23/24, 3:33 PM
by denz88 on 7/23/24, 3:02 PM
by chown on 7/23/24, 4:10 PM
On a related note, for those interested in experimenting with large language models locally, I've been working on an app called Msty [1]. It allows you to run models like this with just one click and features a clean, functional interface. Just added support for both 8B and 70B. Still in development, but I'd appreciate any feedback.
[1]: https://msty.app
by zhanghsfz on 7/23/24, 7:02 PM
Let us know if you have other needs!
by TechDebtDevin on 7/23/24, 2:49 PM
by ChrisArchitect on 7/23/24, 4:21 PM
Open Source AI Is the Path Forward
https://about.fb.com/news/2024/07/open-source-ai-is-the-path...
by Atreiden on 7/23/24, 3:03 PM
Seems like the biggest GPU node they have is the p5.48xlarge @ 640GB (8xH100s). Routing between multiple nodes would be too slow unless there's an InfiniBand fabric you can leverage. Interested to know if anyone else is exploring this.
by TheAceOfHearts on 7/23/24, 3:04 PM
by diimdeep on 7/23/24, 3:17 PM
by rcarmo on 7/23/24, 7:49 PM
by bick_nyers on 7/23/24, 7:29 PM
by jiriro on 7/24/24, 8:07 PM
And answer queries like:
Give all <myObject> which refer to <location> which refer to an Indo-European <language>.
by albert_e on 7/23/24, 3:00 PM
https://github.com/meta-llama/llama-models/blob/main/models/...
by IceHegel on 7/23/24, 7:27 PM
by breadsniffer on 7/23/24, 7:12 PM
by daft_pink on 7/23/24, 3:19 PM
by yinser on 7/23/24, 3:07 PM
by casper14 on 7/23/24, 2:53 PM
by htk on 7/23/24, 11:44 PM
by Vagantem on 7/23/24, 3:01 PM
by kristianp on 7/23/24, 9:00 PM
by ofermend on 7/23/24, 11:07 PM
by zhanghsfz on 7/23/24, 7:02 PM
Would love to hear your feedback!
by ThrowawayTestr on 7/23/24, 2:57 PM
by Jiahang on 7/24/24, 12:52 PM
by stiltzkin on 7/23/24, 8:06 PM
by hubraumhugo on 7/23/24, 2:58 PM
Meta's goal from the start was to target OpenAI and the other proprietary model players with a "scorched earth" approach by releasing powerful open models to disrupt the competitive landscape.
Meta can likely outspend any other AI lab on compute and talent:
- OpenAI makes an estimated revenue of $2B and is likely unprofitable. Meta generated a revenue of $134B and profits of $39B in 2023.
- Meta's compute resources likely outrank OpenAI by now.
- Open source likely attracts better talent and researchers.
- One possible outcome could be the acquisition of OpenAI by Microsoft to catch up with Meta.
The big winners of this: devs and AI product startups