Trinity large: An open 400B sparse MoE model
Posted by linolevan 22 hours ago
Comments
Comment by mynti 15 hours ago
Comment by zamadatix 1 hour ago
E.g. gemini-3-pro tops the lmarena text chart today at 1488 vs 1346 for gpt-4o-2024-05-13. That's a win rate of 70% (where 50% is equal chance of winning) over 1.5 years. Meanwhile, even the open weights stuff OpenAI gave away last summer scores between the two.
The exception seems to be net new benchmarks/benchmark versions. These start out low and then either quickly get saturated or hit a similar wall after a while.
Comment by gwern 6 minutes ago
Why do you care about LM Arena? It has so many problems, and the fact that no one would suggest using GPT-4o for doing math or coding right now, or much of anything, should tell you that a 'win rate of 70%' does not mean whatever it looks like it means. (Does GPT-4o solve roughly as many Erdos questions as gemini-3-pro...? Can you write roughly as good poetry?)
Comment by linolevan 22 hours ago
Comment by mwcampbell 1 hour ago
Comment by timschmidt 1 hour ago
That said, there are folks out there doing it. https://github.com/lyogavin/airllm is one example.
Comment by antirez 1 hour ago
Comment by notpublic 27 minutes ago
https://frame.work/products/desktop-diy-amd-aimax300/configu...
Comment by Scipio_Afri 13 minutes ago
Comment by notpublic 1 minute ago
Comment by syntaxing 16 minutes ago
Comment by Alifatisk 32 minutes ago
Also, why are they comparing with Llama 4 Maverick? Wasn’t it a flop?
Comment by QuadmasterXLII 13 minutes ago
Comment by frogperson 1 hour ago
Comment by someotherperson 1 hour ago
Comment by jetpackjoe 43 minutes ago
Comment by greggh 1 hour ago
Comment by observationist 6 hours ago