This repository has been archived by the owner on Aug 13, 2024. It is now read-only.
Small Models and In Context Learning #4
DataBassGit
started this conversation in
General
Replies: 1 comment 1 reply
-
Yeah that's the idea. I've heard rumors that Gemini will have "up to the second" updates, so either it's got massive in context learning or some other way to slipstream information in at reference |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The problem with large corporate language models is that they become out of date very quickly. You also have the issue of large models getting out of date. For example, a python library gets an update and old code is no longer valid. It costs a lot of money to update that model, and the maintainers are not going to update it every time a new update comes out for a python module.
Therefore, in context learning has to be used in order to leverage knowledge that was created after the models training date. (2021 for GPT-4) I.e. you inject a KB article into the prompt for the python module you want to write code with. This approach can be expanded to be used for all implementations of language modules.
Therefore, you don't need 180b parameter modules to do good NLP processing. With a functional and efficient prompt enrichment protocol, small models with enriched prompts can accomplish most of the tasks that large models can. (As long as the context length and attention mechanisms can handle the enriched prompt.) This could allow us to use smaller locally hosted modules that could potentially run faster and on a single GPU for many of our prompt executions, then leverage larger corporate modules only for tasks where they are needed and excel at.
Beta Was this translation helpful? Give feedback.
All reactions