-
-
Notifications
You must be signed in to change notification settings - Fork 3.4k
Add support for SageMaker Inference Components in sagemaker chat #10603
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
@bobbywlindsey sagemaker_chat no longer uses the openai_like/ flow. It's been migrated to our common base_llm_http_handler.py Can you please update your PR to reflect the change? You should just need to make any mods here -
|
@krrishdholakia Nice refactor! I put my changes where you suggested and seems to work nicely 👍🏻 Could you review again? Thanks! |
Hey @bobbywlindsey could you please add some unit testing inside tests/litellm? |
Hey @ishaan-jaff, any chance you could take a look at this PR? Thanks! |
@krrishdholakia I see you're the reviewer for this PR - all the checks have passed and there are no conflict with base. Can we merge? Thank you! |
Title
Add support for SageMaker Inference Components in sagemaker chat
Relevant issues
Fixes #9909
Pre-Submission checklist
Please complete all items before asking a LiteLLM maintainer to review your PR
tests/litellm/
directory, Adding at least 1 test is a hard requirement - see detailsmake test-unit
Type
🆕 New Feature
🐛 Bug Fix
Changes
If
model_id
is present as a parameter forcompletion(model="sagemaker_chat/*"...)
calls:model_id
as the Inference Componentmodel
key and value from body of request