Nel Nibcord
|
d69a9c4d43
Enabled inference engine intercompatibility
|
8 months ago |
Alex Cheema
|
e9ba815c21
add qwen2.5 coder 3b,14b,32b
|
8 months ago |
Austin
|
5435671cd9
Add 32b Qwen 2.5
|
8 months ago |
Alex Cheema
|
858421a3a7
Merge pull request #418 from BatSmacker84/llama-3.2-support
|
8 months ago |
Alex Cheema
|
98948441e3
add llama 3.1 405b 8bit at mlx-community/Meta-Llama-3.1-405B-Instruct-8bit
|
8 months ago |
Alex Cheema
|
fcaebd3b50
add Gemma2 9b and Gemma2 27bg
|
8 months ago |
Ogden Wells
|
029dc5f8a7
added new model info for 1B and 3B model sizes
|
8 months ago |
Alex Cheema
|
36c1f68cc7
update llama-3.1-405b-8bit model id to IntuitIntel/Meta-Llama-3.1-405B-Instruct-8bit
|
8 months ago |
Alex Cheema
|
c8438b6d23
add llama-3.1-405b-8bit
|
8 months ago |
Alex Cheema
|
6ae6ebeb05
revert back to CORRECTED
|
8 months ago |
Alex Cheema
|
e8e05e15a1
remove CORRECTED llama 70b
|
8 months ago |
Alex Cheema
|
436709e509
revert back to CORRECTED
|
8 months ago |
Alex Cheema
|
8d524bfe71
change mlx-community/Meta-Llama-3.1-70B-Instruct-bf16-CORRECTED to mlx-community/Meta-Llama-3.1-70B-Instruct-bf16 since it works now
|
8 months ago |
Alex Cheema
|
98ea71edda
run format.py on ./exo
|
8 months ago |
Alex Cheema
|
9b8d58c464
fix dummy setup
|
8 months ago |
Alex Cheema
|
82a708f974
rm ministral-8b
|
9 months ago |
Alex Cheema
|
6bd0f07a32
add ministral-8b
|
9 months ago |
Alex Cheema
|
61ee67c955
add nemotron-70b and nemotron-70b-bf16
|
9 months ago |
Mark Van Aken
|
82c7ce69f6
Point `llama-3.1-70b-bf16` model to the actually bf16 version
|
9 months ago |
Alex Cheema
|
abca3bfa37
add support for qwen2.5 coder 1.5b and 7b
|
9 months ago |
Alex Cheema
|
777102c997
add support for llama 3.2
|
9 months ago |
Alex Cheema
|
835e209791
add deepseek-coder-v2.5
|
10 months ago |
Alex Cheema
|
dee83e4840
add more qwen2.5 models: mlx-community/Qwen2.5-7B-Instruct-4bit mlx-community/Qwen2.5-Math-7B-Instruct-4bit mlx-community/Qwen2.5-72B-Instruct-4bit mlx-community/Qwen2.5-Math-72B-Instruct-4bit
|
10 months ago |
Alex Cheema
|
3597fba356
add support for qwen2.5, initially adding mlx-community/Qwen2.5-14B-Instruct-4bit
|
10 months ago |
Alex Cheema
|
2948a83448
add llama-3.1-70b-bf16 model option
|
10 months ago |
Alex Cheema
|
dc3b2bde39
use NousResearch/Meta-Llama-3.1-70B-Instruct as tinygrad llama-3.1-70b model, previously using non-instruct model
|
10 months ago |
Alex Cheema
|
f53056dede
more compact operator formatting
|
11 months ago |
Alex Cheema
|
ea70c9fb76
reformat with yapf format.py
|
11 months ago |
Alex Cheema
|
e84304317c
add a cli that can be triggered with --run-model <model> --prompt <prompt>
|
11 months ago |