Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

NVIDIA NIM LLM Hosting Pattern #560

Closed
hustshawn opened this issue Jun 20, 2024 · 2 comments
Closed

NVIDIA NIM LLM Hosting Pattern #560

hustshawn opened this issue Jun 20, 2024 · 2 comments
Labels
gen-ai pattern Distributed Training and Inference Patterns for Various Generative AI Large Language Models (LLMs) good first issue Good for newcomers

Comments

@hustshawn
Copy link
Contributor

hustshawn commented Jun 20, 2024

Community Note

  • Please vote on this issue by adding a 👍 reaction to the original issue to help the community and maintainers prioritize this request
  • Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
  • If you are interested in working on this issue or have submitted a pull request, please leave a comment

What is the outcome that you are trying to reach?

NVIDIA NIM pvovides an easy way to self-host LLMs with containers. AWS can provide series of solutions to help customer to deploy the NIM provided LLMs in a performant and cost optimized way with EKS, Karpenter, Spot etc.
Would like to have a pattern for the customer to easily deploy the NIM LLMs in AWS.

Describe the solution you would like

Provide a IaC pattern in Data on EKS to showcase deploy NVIDIA NIM provided LLMs. Will use Karpenter, Spot, and Bottlerocket with image caching techniques which I introduced in this AWS blog.

Describe alternatives you have considered

Additional context

NVIDIA NIM: https://www.nvidia.com/en-us/ai/
Features and architecture: https://docs.nvidia.com/nim/large-language-models/latest/introduction.html

@vara-bonthu
Copy link
Collaborator

This will be a great addition.

Step1: Leverage Existing Terraform Blueprint:
Utilize the existing Terraform blueprint to set up infrastructure components like EKS Cluster, Karpenter, and other addons. Update this template if required specific to Bottlerocket implementation etc.

Step2: Add a New Folder:
Create a new folder under gen-ai named nvidia-nims.
Place all the necessary code for your blueprint in this folder.

Step3: Create Documentation:
Add a new documentation file under the Website docs with the naming convention nvidia-nim-<modelname>.

@vara-bonthu vara-bonthu added gen-ai pattern Distributed Training and Inference Patterns for Various Generative AI Large Language Models (LLMs) good first issue Good for newcomers labels Jun 20, 2024
@hustshawn
Copy link
Contributor Author

Thank you @vara-bonthu . I am gonna implement it, please feel free to assign this to me. I raise a PR once it is finished.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
gen-ai pattern Distributed Training and Inference Patterns for Various Generative AI Large Language Models (LLMs) good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

3 participants