Codesota · Models · Phi-3.5-mini-instructmicrosoft13 results · 2 benchmarks
Model card

Phi-3.5-mini-instruct.

microsoftopen-source3.82B params
§ 01 · Benchmarks

Every benchmark Phi-3.5-mini-instruct has a recorded score for.

#BenchmarkArea · TaskMetricValueRankDateSource
01CPTU-BenchNatural Language Processing · Polish Text Understandingphraseology2.4%#79/93source ↗
02CPTU-BenchNatural Language Processing · Polish Text Understandingtricky-questions1.0%#82/93source ↗
03CPTU-BenchNatural Language Processing · Polish Text Understandingaverage2.0%#84/93source ↗
04CPTU-BenchNatural Language Processing · Polish Text Understandinglanguage-understanding2.1%#86/93source ↗
05CPTU-BenchNatural Language Processing · Polish Text Understandingsentiment2.4%#88/93source ↗
06Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generaleq-bench17.0%#216/299source ↗
07Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalaverage37.5%#247/491source ↗
08Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalppc67.1%#261/490source ↗
09Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generaldyk43.7%#287/489source ↗
10Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalpolemo2-in69.0%#366/490source ↗
11Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalbelebele62.8%#367/490source ↗
12Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalpolqa-open-book64.8%#441/489source ↗
13Open PL LLM LeaderboardNatural Language Processing · Polish LLM Generalcbd13.8%#480/490source ↗
Rank column shows this model’s position vs all other models scored on the same benchmark + metric (competitors after the slash). #1 in red means current SOTA. Sorted by rank, then newest result.
§ 02 · Strengths by area

Where Phi-3.5-mini-instruct actually performs.

Natural Language Processing
2
benchmarks
avg rank #237.2
§ 04 · Related models

Other microsoft models scored on Codesota.

Phi-3-medium-4k-instruct
0 results
Phi-3-mini-4k-instruct
0 results
Phi-3-small-8k-instruct
0 results
Phi-3.5-MoE-instruct
0 results
Phi-4-mini-instruct
3.84B params · 0 results
phi-4
14.7B params · 0 results
🚧WizardLM-2-7B
0 results
§ 05 · Sources & freshness

Where these numbers come from.

speakleash/open_pl_llm_leaderboard
8
results
SpeakLeash/CPTU-Bench
5
results
13 of 13 rows marked verified.