-
Notifications
You must be signed in to change notification settings - Fork 596
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]: Integration with Amazon bedrock issue #1076
Comments
Yes that's right, though it's possible to adjust this, we just haven't tested. |
You ok for me to attempt this ? Making the request from the 'right' account :). Thanks, |
@malphi - I was trying to reproduce the error , so while we cannot use the other models apart from the anthropic.claude-v1/2 ones , I don't get the error as if one of the other models is selected the code defaults to anthropic.claude-v2. And based on the error , may be you don't have access the "anthropic.claude-v2" model where you were testing. The request for including other models is still valid , just that the response would not be an error , but a attempt at claude. |
@AlexsJones - Have made an attempt and submitted a pull #1099 not too sure about the process so looking for guidance on the missing bits , added a signature on new pushes with no changes but intermediate pushes still have a error. We should not need additional documentation as the documents don't outline which models are supported specifically, so current instructions work. There are next steps to add all the other models, but did not want to make too many additions on a single pr, can add other models as a follow-up. |
Thanks @awsyshah ill take a look at that |
@AlexsJones cleaned it up and submitted a fresh PR. |
Yes, It works when I fix model permission issue, Thanks. |
Thanks for your support. I went through the PR, my suggestion is, can you provide more flexible way to add models such as configuration? Otherwise we have to change code if the AI provider add new models. |
I agree, it would be good to be able to pass the models as a parameter, but the nature of the models with different request & response formats , different max limits. my thinking is it will still need some significant rework , potentially a local conf file for the user to define the format of the request and response , which would be a total revamp of the current solution approach. I might suggest we extend the client for now with the current approach and can look at a wider PR to completely parameterize the model input. @AlexsJones thoughts ? |
I hear what you're saying. But by restricting the models we support, we are intentionally putting some experience guard rails up. We know that they each take slightly different parameterisation. I think the small compromise here is needing a PR to support more, but it keeps the quality of the experience high for everyone. |
@AlexsJones Let me know your thoughts on the PR if it looks good now with the modifications. |
Checklist
Affected Components
K8sGPT Version
0.3.29
Kubernetes Version
v1.29
Host OS and its Version
MacOS
Steps to reproduce
command is : k8sgpt auth add --backend amazonbedrock --model ai21.j2-ultra-v1 --providerRegion us-east-1
Error: failed while calling AI provider amazonbedrock: AccessDeniedException: You don't have access to the model with the specified model ID.
.Expected behaviour
integration with bedrock success.
Actual behaviour
Error: failed while calling AI provider amazonbedrock: AccessDeniedException: You don't have access to the model with the specified model ID.
Additional Information
No response
The text was updated successfully, but these errors were encountered: