-
-
Notifications
You must be signed in to change notification settings - Fork 359
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A model that picks the right sized model #32
Comments
This may be a template and not a model - perhaps Not sure what the YAML would look like. A model might be better through, since then you could combine a template with the |
I think it's a special model called How should it handle some users not having GPT-4 32k access? I think it should try anyway and error if they don't have the model - it would have errored anyway since they were over 32k tokens. |
Also need to consider the 3.5 4k vs 16k, guessing this is going to be a pattern that continues as well; models that are "the same" but differ in context length (and pricing). I think there needs to be some concept of "flavors" of models, and in |
Worth noting this is a problem that other tools are facing right now as well. I'm not aware of any consensus on how to handle it as of yet, but it's probably worth looking into. |
this is relevant to the new |
Dropped from the 0.5 milestone, it's not critical for that. I'm actually thinking this might make more sense as a |
Count tokens with tiktoken and switch to the 16k or 32k models if necessary.
The text was updated successfully, but these errors were encountered: