diff --git a/README.md b/README.md index 48b77323..fddb7421 100644 --- a/README.md +++ b/README.md @@ -76,6 +76,10 @@ Here are some example models that can be downloaded: | Model | Parameters | Size | Download | | ------------------ | ---------- | ----- | -------------------------------- | +| Gemma 3 | 1B | 815MB | `ollama run gemma3:1b` | +| Gemma 3 | 4B | 3.3GB | `ollama run gemma3` | +| Gemma 3 | 12B | 8.1GB | `ollama run gemma3:12b` | +| Gemma 3 | 27B | 17GB | `ollama run gemma3:27b` | | QwQ | 32B | 20GB | `ollama run qwq` | | DeepSeek-R1 | 7B | 4.7GB | `ollama run deepseek-r1` | | DeepSeek-R1 | 671B | 404GB | `ollama run deepseek-r1:671b` | @@ -88,9 +92,6 @@ Here are some example models that can be downloaded: | Llama 3.1 | 405B | 231GB | `ollama run llama3.1:405b` | | Phi 4 | 14B | 9.1GB | `ollama run phi4` | | Phi 4 Mini | 3.8B | 2.5GB | `ollama run phi4-mini` | -| Gemma 2 | 2B | 1.6GB | `ollama run gemma2:2b` | -| Gemma 2 | 9B | 5.5GB | `ollama run gemma2` | -| Gemma 2 | 27B | 16GB | `ollama run gemma2:27b` | | Mistral | 7B | 4.1GB | `ollama run mistral` | | Moondream 2 | 1.4B | 829MB | `ollama run moondream` | | Neural Chat | 7B | 4.1GB | `ollama run neural-chat` | diff --git a/cmd/interactive.go b/cmd/interactive.go index af58f4fc..7c11ab83 100644 --- a/cmd/interactive.go +++ b/cmd/interactive.go @@ -195,6 +195,10 @@ func generateInteractive(cmd *cobra.Command, opts runOptions) error { opts.Messages = []api.Message{} fmt.Printf("Loading model '%s'\n", opts.Model) if err := loadOrUnloadModel(cmd, &opts); err != nil { + if strings.Contains(err.Error(), "not found") { + fmt.Printf("error: %v\n", err) + continue + } return err } continue