ollama: ship a bunch of new models
This commit is contained in:
@@ -40,11 +40,17 @@ let
|
||||
deepseek-r1-abliterated-14b
|
||||
deepseek-r1-abliterated-32b
|
||||
deepseek-r1-abliterated-70b
|
||||
dolphin-mistral-7b # UNCENSORED mistral; compliant
|
||||
dolphin-mixtral-8x7b # about as fast as a 14b model, similar quality results. uncensored, but still preachy
|
||||
devstral-24b
|
||||
dolphin3-8b
|
||||
# dolphin-mistral-7b # UNCENSORED mistral; compliant
|
||||
# dolphin-mixtral-8x7b # about as fast as a 14b model, similar quality results. uncensored, but still preachy
|
||||
# falcon2-11b # code examples are lacking
|
||||
# gemma2-9b # fast, but not great for code
|
||||
gemma2-27b # generates at 1word/sec, but decent coding results if you can wrangle it
|
||||
# gemma2-27b # generates at 1word/sec, but decent coding results if you can wrangle it
|
||||
gemma3-12b
|
||||
gemma3-27b
|
||||
gemma3n-e2b
|
||||
gemma3n-e4b
|
||||
# glm4-9b # it generates invalid code
|
||||
# hermes3-8b # FAST, but unwieldy
|
||||
# llama3-chatqa-8b # it gets stuck
|
||||
@@ -54,15 +60,25 @@ let
|
||||
# llama3_3-70b # non-compliant; dodges iffy questions
|
||||
llama3_3-abliterated-70b # compliant, but slower and not as helpful as deepseek-r1-abliterated-70b
|
||||
magicoder-7b # it generates valid, if sparse, code
|
||||
magistral-24b
|
||||
marco-o1-7b # untested
|
||||
# mistral-7b # it generates invalid code
|
||||
# mistral-nemo-12b # it generates invalid code
|
||||
mistral-small-22b # quality comparable to qwen2_5
|
||||
# mistral-small-22b # quality comparable to qwen2_5
|
||||
mistral-small3_2-24b
|
||||
# mistral-large-123b # times out launch on desko
|
||||
# mixtral-8x7b # generates valid, if sparse, code; only for the most popular languages
|
||||
olmo2-13b
|
||||
openthinker-7b
|
||||
openthinker-32b
|
||||
orca-mini-7b
|
||||
# phi3_5-3b # generates invalid code
|
||||
phi4-14b
|
||||
# qwen2_5-7b # notably less quality than 32b (i.e. generates invalid code)
|
||||
qwen2_5-14b # *almost* same quality to 32b variant, but faster
|
||||
# qwen2_5-14b # *almost* same quality to 32b variant, but faster
|
||||
qwen3-8b
|
||||
qwen3-14b
|
||||
qwen3-30b
|
||||
# qwen2_5-32b-instruct-q2_K # lower-res version of default 32b (so, slightly faster, but generates invalid code where the full res generates valid code)
|
||||
qwen2_5-32b # generates 3~5 words/sec, but notably more accurate than coder-7b
|
||||
qwen2_5-abliterate-7b
|
||||
|
Reference in New Issue
Block a user