Will a Mamba 7b model trained on 2 trillion tokens outperform Llama2-13B
21
Ṁ738
Jul 1
66%
chance

Question will resolve positive if someone trains a Mamba (https://twitter.com/tri_dao/status/1731728602230890895) language model with <=7.5billion parameters on <=2 trillion tokens that outperforms Llama2-13B on the huggingface open llm leaderboard (https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)

Get Ṁ1,000 play money
Sort by: