Skip to content

Commit

Permalink
Change server port for weather agent
Browse files Browse the repository at this point in the history
  • Loading branch information
ehsanmok committed Feb 18, 2025
1 parent f565ec2 commit a0af3f3
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
4 changes: 2 additions & 2 deletions ai-weather-agent/Procfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
max-serve-llm: (magic global install max-pipelines && magic global update max-pipelines) || true; MAX_SERVE_PORT=8000 MAX_SERVE_HOST=localhost HUGGING_FACE_HUB_TOKEN=$(cat backend/.env | grep HUGGING_FACE_HUB_TOKEN | cut -d '=' -f2) max-pipelines serve --huggingface-repo-id=modularai/Llama-3.1-8B-Instruct-GGUF --max-length 2048
max-serve-embedding: (magic global install max-pipelines && magic global update max-pipelines) || true; MAX_SERVE_PORT=7999 MAX_SERVE_HOST=localhost HUGGING_FACE_HUB_TOKEN=$(cat backend/.env | grep HUGGING_FACE_HUB_TOKEN | cut -d '=' -f2) max-pipelines serve --huggingface-repo-id=sentence-transformers/all-mpnet-base-v2
max-serve-llm: (magic global install max-pipelines && magic global update max-pipelines) || true; MAX_SERVE_PORT=8010 MAX_SERVE_HOST=0.0.0.0 HUGGING_FACE_HUB_TOKEN=$(cat backend/.env | grep HUGGING_FACE_HUB_TOKEN | cut -d '=' -f2) max-pipelines serve --huggingface-repo-id=modularai/Llama-3.1-8B-Instruct-GGUF --max-length 2048
max-serve-embedding: (magic global install max-pipelines && magic global update max-pipelines) || true; MAX_SERVE_PORT=7999 MAX_SERVE_HOST=0.0.0.0 HUGGING_FACE_HUB_TOKEN=$(cat backend/.env | grep HUGGING_FACE_HUB_TOKEN | cut -d '=' -f2) max-pipelines serve --huggingface-repo-id=sentence-transformers/all-mpnet-base-v2
backend: cd backend && magic run backend
frontend: cd frontend && magic run frontend
2 changes: 1 addition & 1 deletion ai-weather-agent/backend/src/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ async def lifespan(app: FastAPI):

CACHE_TTL = 3600 # 1 hour

LLM_SERVER_URL = os.getenv("LLM_SERVER_URL", "http://0.0.0.0:8000/v1")
LLM_SERVER_URL = os.getenv("LLM_SERVER_URL", "http://0.0.0.0:8010/v1")
EMBEDDING_SERVER_URL = os.getenv("EMBEDDING_SERVER_URL", "http://0.0.0.0:7999/v1")
EMBEDDING_MODEL = os.getenv(
"EMBEDDING_MODEL", "sentence-transformers/all-mpnet-base-v2"
Expand Down

0 comments on commit a0af3f3

Please sign in to comment.