ollamaPackages.{qwen2_5-7b,qwen2_5-14b}: init, and cleanup the surroundings
This commit is contained in:
@@ -19,12 +19,14 @@ let
|
||||
modelSources = pkgs.symlinkJoin {
|
||||
name = "ollama-models";
|
||||
paths = with pkgs.ollamaPackages; [
|
||||
llama3_1-70b
|
||||
# llama3_1-70b # generates like 1 word/sec, decent output (comparable to qwen2_5-32b)
|
||||
llama3_2-3b
|
||||
mistral
|
||||
mistral-large
|
||||
mixtral-8x7b
|
||||
qwen2_5-32b
|
||||
qwen2_5-7b
|
||||
qwen2_5-14b
|
||||
qwen2_5-32b # generates 3~5 words/sec, but notably more accurate than coder-7b
|
||||
qwen2_5-coder-7b
|
||||
];
|
||||
};
|
||||
|
Reference in New Issue
Block a user