Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

change model #13

Open
slowmagic10 opened this issue Jan 2, 2025 · 1 comment
Open

change model #13

slowmagic10 opened this issue Jan 2, 2025 · 1 comment

Comments

@slowmagic10
Copy link

can I use other nim models like llama-3.1 8b instead of 70b?

@omihub777
Copy link

@slowmagic10
Yes, you can

In case of using nvidia-hosted LLMs, you just need to set an environment variable by export APP_LLM_MODELNAME=meta/llama-3.1-8b-instruct before run docker compose up

For the self-hosted deployment, you additionally need to change the image on deploy/compose/nims.yaml:

services:
  nemollm-inference:
    container_name: nemollm-inference-microservice
    # image: nvcr.io/nim/meta/llama-3.1-70b-instruct:1.1 # comment out
   image: nvcr.io/nim/meta/llama-3.1-8b-instruct:latest # use 8b instead

Be aware that llama-3.1-8b might not be performant enough to run this blueprint successfully

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants