diff --git a/packages/backend/src/assets/llama-stack-images.json b/packages/backend/src/assets/llama-stack-images.json index 0bb7ad9e1..e5a890550 100644 --- a/packages/backend/src/assets/llama-stack-images.json +++ b/packages/backend/src/assets/llama-stack-images.json @@ -1,3 +1,3 @@ { - "default": "ghcr.io/containers/podman-ai-lab-stack:8d6a4a9a7c587c0a8e44703dd750355256e7a796" + "default": "ghcr.io/containers/podman-ai-lab-stack:a06f399ebf7cb2645af126da0e84395db9bb0d1a" } diff --git a/packages/backend/src/managers/llama-stack/llamaStackManager.spec.ts b/packages/backend/src/managers/llama-stack/llamaStackManager.spec.ts index 4c7893019..9e6646eb7 100644 --- a/packages/backend/src/managers/llama-stack/llamaStackManager.spec.ts +++ b/packages/backend/src/managers/llama-stack/llamaStackManager.spec.ts @@ -308,6 +308,8 @@ test('requestcreateLlamaStackContainerss registers all local models', async () = 'models', 'register', 'Model 1', + '--provider-id', + 'podman-ai-lab', ]); expect(podmanConnection.execute).toHaveBeenCalledWith(expect.anything(), [ 'exec', @@ -316,6 +318,8 @@ test('requestcreateLlamaStackContainerss registers all local models', async () = 'models', 'register', 'Model 2', + '--provider-id', + 'podman-ai-lab', ]); }); diff --git a/packages/backend/src/managers/llama-stack/llamaStackManager.ts b/packages/backend/src/managers/llama-stack/llamaStackManager.ts index 86bc89bf6..43cc8f6e5 100644 --- a/packages/backend/src/managers/llama-stack/llamaStackManager.ts +++ b/packages/backend/src/managers/llama-stack/llamaStackManager.ts @@ -501,6 +501,8 @@ export class LlamaStackManager implements Disposable { 'models', 'register', model.name, + '--provider-id', + 'podman-ai-lab', ]); }, );