Skip to content

Commit 627949d

Browse files
authored
Swap old Llama model with Phi-3 (#1666)
1 parent 40c293d commit 627949d

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -97,12 +97,12 @@ Every model is written from scratch to maximize performance and remove layers of
9797
| Model | Model size | Author | Reference |
9898
|----|----|----|----|
9999
| Llama 3 & 3.1 | 8B, 70B, 405B | Meta AI | [Meta AI 2024](https://github.com/meta-llama/llama3) |
100-
| Llama 2 | 7B, 13B, 70B | Meta AI | [Touvron et al. 2023](https://arxiv.org/abs/2307.09288) |
101100
| Code Llama | 7B, 13B, 34B, 70B | Meta AI | [Rozière et al. 2023](https://arxiv.org/abs/2308.12950) |
102101
| Mixtral MoE | 8x7B | Mistral AI | [Mistral AI 2023](https://mistral.ai/news/mixtral-of-experts/) |
103102
| Mistral | 7B | Mistral AI | [Mistral AI 2023](https://mistral.ai/news/announcing-mistral-7b/) |
104103
| CodeGemma | 7B | Google | [Google Team, Google Deepmind](https://ai.google.dev/gemma/docs/codegemma) |
105104
| Gemma 2 | 2B, 9B, 27B | Google | [Google Team, Google Deepmind](https://storage.googleapis.com/deepmind-media/gemma/gemma-2-report.pdf) |
105+
| Phi 3 | 3.8B | Microsoft | [Abdin et al. 2024](https://arxiv.org/abs/2404.14219) |
106106
| ... | ... | ... | ... |
107107

108108
<details>

0 commit comments

Comments
 (0)