Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

host_ip causes an error in ChatQnA AIPC deployment with Ollama #703

Open
1 of 6 tasks
devpramod opened this issue Aug 30, 2024 · 2 comments
Open
1 of 6 tasks

host_ip causes an error in ChatQnA AIPC deployment with Ollama #703

devpramod opened this issue Aug 30, 2024 · 2 comments
Assignees

Comments

@devpramod
Copy link
Collaborator

devpramod commented Aug 30, 2024

Priority

P2-High

OS type

Ubuntu

Hardware type

AI-PC

Installation method

  • Pull docker images from hub.docker.com
  • Build docker images from source

Deploy method

  • Docker compose
  • Docker
  • Kubernetes
  • Helm

Running nodes

Single Node

What's the version?

25174c0

Description

In GenAIComps, ollama is tested with localhost:
curl http://localhost:11434/api/generate -d '{ "model": "llama3", "prompt":"Why is the sky blue?" }'
`
curl http://127.0.0.1:9000/v1/chat/completions -X POST -d '{"model": "llama3", "query":"What is Deep Learning?","max_new_tokens":32,"top_k":10,"top_p":0.95,"typical_p":0.95,"temperature":0.01,"repetition_penalty":1.03,"streaming":true}' -H 'Content-Type: application/json'

`
which works fine.
But in GenAIExamples - instead of localhost, host_ip is used which causes an error (Connection refused)

Reproduce steps

cd GenAIExamples/ChatQnA/docker/aipc

docker compose up -d

run ollama

curl http://${host_ip}:9000/v1/chat/completions
-X POST
-d '{"query":"What is Deep Learning?","max_new_tokens":17,"top_k":10,"top_p":0.95,"typical_p":0.95,"temperature":0.01,"repetition_penalty":1.03,"streaming":true}'
-H 'Content-Type: application/json'

Raw log

[No response](curl: (7) Failed to connect to x.x.x.x port 11434 after 0 ms: Connection refused)

@dcmiddle
Copy link
Contributor

@devpramod can you add a meaningful title to the issue?
Do you have host_ip set?

@devpramod devpramod changed the title [Bug] host_ip causes an error in ChatQnA AIPC deployment Sep 13, 2024
@devpramod devpramod changed the title host_ip causes an error in ChatQnA AIPC deployment host_ip causes an error in ChatQnA AIPC deployment with Ollama Sep 13, 2024
@devpramod
Copy link
Collaborator Author

@dcmiddle Yes all other services work fine with host_ip
For Ollama, both the backend LLM service (i.e. Ollama itself) and the LLM microservice don't work with host_ip, need to set to localhost

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants