-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
.Net: New Feature: .NET support calling Inference profiles with AWS Bedrock #10047
Comments
@RogerBarreto can you take a look at this? |
Is this why I had to modify the To use Meta Llama 3.1 8B instruct, I have to use the modelId
With your proposal, we would use the foundation model's ID to create the kernel instance with the inference profile Id. Because AWS uses the inference profile ID as the model ID in the API request, the internal code will check and use the inference profile ID as the request's model ID. I understand it would support the user-defined inference profile ID and still support the semantic kernel's Amazon Bendrock library to create the appropriate foundation model service. Is this a correct understanding? |
i'm also encountering this issue |
Name: Feature request .NET support calling Inference profiles with AWS Bedrock
About: For tracking and billing purposes using inference profiles with AWS Bedrock allows usage attribution to a specific model across multiple teams within an organization.
Problem
Currently Semantic Kernel expects
modelId
to match the formatprovider.modelName
This is used to determine the client instantiated (e.g. meta, amazon, etc.) ThemodelId
is also used to build the request.When using an inference profile, the inference profile replaces the
modelId
in the request. By allowing the usage ofinferenceProfile
, multiple teams can independently track and attribute usage when working with the same model.Current Implementation of service selection: BedrockServiceFactory.cs
Proposal
Create a new parameter to track inference profile e.g.
inferenceProfile
When building the request to bedrock, if the
inferenceProfile
is present use it instead of the foundationalmodelId
when generating the request Uri, otherwise use the foundationalmodelId
.Here are some example suggestions:
Or if the
InferenceProfile
was allowed on thePromptExecutionSettings
then:Is mise, le meas, Matthew
The text was updated successfully, but these errors were encountered: