diff --git a/bucket/ollama.json b/bucket/ollama.json index 12868deda8375..a712951741285 100644 --- a/bucket/ollama.json +++ b/bucket/ollama.json @@ -3,23 +3,14 @@ "description": "Get up and running with large language models locally.", "homepage": "https://ollama.com/", "license": "MIT", - "url": "https://github.com/ollama/ollama/releases/download/v0.1.44/OllamaSetup.exe", - "hash": "640c587550810b941e384618e4a0130e3cda56ac67f7894b821c141aefd6ad0c", - "innosetup": true, + "url": "https://github.com/ollama/ollama/releases/download/v0.1.44/ollama-windows-amd64.zip", + "hash": "483fef57f094f5a52d865182a0d6f9a2e8a59eb44dd756b4517ea9b747d90338", "bin": "ollama.exe", - "shortcuts": [ - [ - "ollama app.exe", - "Ollama", - "", - "app.ico" - ] - ], "checkver": { "github": "https://github.com/ollama/ollama" }, "autoupdate": { - "url": "https://github.com/ollama/ollama/releases/download/v$version/OllamaSetup.exe" + "url": "https://github.com/ollama/ollama/releases/download/v$version/ollama-windows-amd64.zip" }, "notes": "You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 13B models, and 32 GB to run the 33B models." }