This build will experience a 4096 memory input too large error with pasting of a large prompt into the UI
ongoing development of Professor Codephreak LLM here
To avoid confusion automindx is the local language model executable and AUTOMINDx is the global deployment autonomous agents development environment
This Professor Codephreak build is being left intact as the first example of the working aGLM prototype
Work continues at automindx as the next evolution of Professor Codephreak's core intention value and success creating AUTOMINDx
"I Am Machine Learning"
project codename = codephreak
project direction = automind automindx
"Professor Codephreak is an expert in machine learning, computer science and computer programming ..."
uiux.py provides Gradio output to local server for local language model interaction
Gradio interacts with html javascript and css custom Gradio
Gradio Javascript Client
Documentation: codephreak = uiux.py + memory.py + automind.py + aglm.py
(uiux.py) provides a user interface using the Gradio library to facilitate user interaction.
This chatbot interface takes the user input, processes it, generating response includingc onversation memory handling storing and managing user instructions and model responses. The result is a local language model prompted to run as "Professor Codephreak is an expert in computer programming ....." that refers to itself as "codephreak"
Contextual Conversation Management
(automind.py) provides the mechanism to format and managing conversation history using the format_to_llama_chat_style function.
(automind.py) creates coherent conversation context integrating memory management and with chatbot behavior.
Handling different model types and initialization is based on model name calling the model from the models folder.
(memory.py) handles Conversation Memory Management by creating a class DialogEntry to represent individual conversation dialog entries.
A function save_conversation_memory is called to save conversation history as JSON files. Memory is created by storing user instructions and model responses in memory files for context management as .json outputs with a timestamp.
(aglm.py)
Initializes the LLAMA language model and tokenizer based on the specified models_folder models handling processing and tokenization of conversation context using the LLAMA model then generating contextually relevant responses using the LLAMA model. AUTOMIND uses a LLAMA language model to generate responses based on user instructions and conversation history. (uiux.py) encompasses user interface design, memory management, context handling, and interaction with the language model to create a conversational experience for users.
To install right click "Save link as ..." automind.install chmod +x automind.install && automind.install
details and verbose procedure
- Right-click the following link: automind.install
- Choose "Save link as..." or "Download linked file" from the context menu.
- Select a location on your computer to save the file.
- from the terminal
- chmod +x automind.install && ./automind.install
python3 uiux.py --model_name="TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML" --tokenizer_name="TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML" --model_type="ggml" --save_history --file_name="llama-2-7b-chat-codeCherryPop.ggmlv3.q4_1.bin"