ollama: ship llama4-16x17b
This commit is contained in:
@@ -59,6 +59,7 @@ let
|
||||
llama3_2-uncensored-3b
|
||||
# llama3_3-70b # non-compliant; dodges iffy questions
|
||||
llama3_3-abliterated-70b # compliant, but slower and not as helpful as deepseek-r1-abliterated-70b
|
||||
llama4-16x17b
|
||||
magicoder-7b # it generates valid, if sparse, code
|
||||
magistral-24b
|
||||
marco-o1-7b # untested
|
||||
|
@@ -78,7 +78,7 @@ ollama API isn't documented anywhere, and it has changed over time, but it's all
|
||||
- released 2025-04-07
|
||||
- <https://ollama.com/library/deepcoder>
|
||||
- <https://pretty-radio-b75.notion.site/DeepCoder-A-Fully-Open-Source-14B-Coder-at-O3-mini-Level-1cf81902c14680b3bee5eb349a512a51>
|
||||
- [ ] llama4 (16x17b, 128x17b)
|
||||
- [x] llama4 (16x17b, 128x17b)
|
||||
- released 2025-04-05
|
||||
- <https://ollama.com/library/llama4>
|
||||
- <https://ai.meta.com/blog/llama-4-multimodal-intelligence/>
|
||||
|
14
pkgs/ollamaPackages/llama4-16x17b/package.nix
Normal file
14
pkgs/ollamaPackages/llama4-16x17b/package.nix
Normal file
@@ -0,0 +1,14 @@
|
||||
# <https://ollama.com/library/llama4>
|
||||
# <https://ai.meta.com/blog/llama-4-multimodal-intelligence/>
|
||||
# released 2025-04-05
|
||||
{ mkOllamaModel }: mkOllamaModel {
|
||||
modelName = "llama4";
|
||||
variant = "16x17b";
|
||||
manifestHash = "sha256-vzFgTiXCXZZOJQvPKKgr+9voivXyNiV/q7J2Kbskx/M=";
|
||||
modelBlob = "9d507a36062c2845dd3bb3e93364e9abc1607118acd8650727a700f72fb126e5";
|
||||
modelBlobHash = "sha256-nVB6NgYsKEXdO7PpM2Tpq8FgcRis2GUHJ6cA9y+xJuU=";
|
||||
paramsBlob = "bee89e20d457c132784e74ae48177b45262ecc7383c085c835ec54da40d2e4e6";
|
||||
paramsBlobHash = "sha256-vuieINRXwTJ4TnSuSBd7RSYuzHODwIXINexU2kDS5OY=";
|
||||
systemBlob = "fc1ffc71ab8ebabe8ec0177ea8ee41d1ea27db856636a517d54324eecdfb3f11";
|
||||
systemBlobHash = "sha256-/B/8cauOur6OwBd+qO5B0eon24VmNqUX1UMk7s37PxE=";
|
||||
}
|
Reference in New Issue
Block a user