Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
singhsidhukuldeepΒ 
posted an update Jun 4
Post
1591
Remember when @Microsoft released Phi-3 models... πŸ€”

Yup, the ones that had πŸ¦™Llama 3 8B beat on MMLU using 3.8B parameters! πŸ†

Now they are on the LMSYS Chatbot Arena Leaderboard! πŸ“ŠπŸ“ˆ

Medium(14B) ranks near GPT-3.5-Turbo-0613, but behind Llama 3 8B. πŸ“‰

Phi-3 Small(7B) is close to Llama-2-70B, and Mistral fine-tunes. πŸ“Š

What about the Phi-3 Mini(3.8B), that was giving Llama 3 8B a run for its money on MMLU? It gets an arena score of 1037 (#73) against 1153 (#22) of Llama 3 8B 🀼

Looks like there is a struggle here between perplexity and inherent knowledge! πŸ€”

And Microsoft picked knowledge with high perplexity 🧠

Now I am even more intrigued: what is @Meta feeding its πŸ¦™ Llamas?🌾

πŸ† Leaderboard: https://chat.lmsys.org/?leaderboard