From 9afe8da9f17bbe2ee6fedb7b286feb624bf0afc7 Mon Sep 17 00:00:00 2001 From: Al-Ekram Elahee Hridoy Date: Mon, 2 Oct 2023 20:36:23 -0600 Subject: [PATCH] Format issue (#69) * Fixed formating issue in the README * Fixed the formating issue in the README * Updated formatting as per review comments * Refactor README.md to highlight use cases and features * Updated README as per feedback * Updated README as per feedback --------- Co-authored-by: Al-Iqram Elahee Co-authored-by: Qingyun Wu --- README.md | 20 ++++++++++++++------ 1 file changed, 14 insertions(+), 6 deletions(-) diff --git a/README.md b/README.md index a8af692a479..1ad3348b925 100644 --- a/README.md +++ b/README.md @@ -59,8 +59,18 @@ For LLM inference configurations, check the [FAQ](https://microsoft.github.io/au ## Quickstart -* Autogen enables the next-gen LLM applications with a generic multi-agent conversation framework. It offers customizable and conversable agents which integrate LLMs, tools, and humans. -By automating chat among multiple capable agents, one can easily make them collectively perform tasks autonomously or with human feedback, including tasks that require using tools via code. For [example](https://github.com/microsoft/autogen/blob/main/test/twoagent.py), +## Multi-Agent Conversation Framework + +Autogen enables the next-gen LLM applications with a generic multi-agent conversation framework. It offers customizable and conversable agents which integrate LLMs, tools, and humans. +By automating chat among multiple capable agents, one can easily make them collectively perform tasks autonomously or with human feedback, including tasks that require using tools via code. + +Features of this use case include: + +- **Multi-agent conversations**: AutoGen agents can communicate with each other to solve tasks. This allows for more complex and sophisticated applications than would be possible with a single LLM. +- **Customization**: AutoGen agents can be customized to meet the specific needs of an application. This includes the ability to choose the LLMs to use, the types of human input to allow, and the tools to employ. +- **Human participation**: AutoGen seamlessly allows human participation. This means that humans can provide input and feedback to the agents as needed. + +For [example](https://github.com/microsoft/autogen/blob/main/test/twoagent.py), ```python from autogen import AssistantAgent, UserProxyAgent, config_list_from_json # Load LLM inference endpoints from an env variable or a file @@ -72,9 +82,6 @@ user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "co user_proxy.initiate_chat(assistant, message="Plot a chart of NVDA and TESLA stock price change YTD.") # This initiates an automated chat between the two agents to solve the task ``` -Multi-agent conversations: AutoGen agents can communicate with each other to solve tasks. This allows for more complex and sophisticated applications than would be possible with a single LLM. -Customization: AutoGen agents can be customized to meet the specific needs of an application. This includes the ability to choose the LLMs to use, the types of human input to allow, and the tools to employ. -Human participation: AutoGen seamlessly allows human participation. This means that humans can provide input and feedback to the agents as needed. This example can be run with ```python @@ -85,8 +92,9 @@ The figure below shows an example conversation flow with AutoGen. ![Agent Chat Example](https://github.com/microsoft/autogen/blob/main/website/static/img/chat_example.png) Please find more [code examples](https://microsoft.github.io/autogen/docs/Examples/AutoGen-AgentChat) for this feature. +## Enhanced LLM Inferences -* Autogen also helps maximize the utility out of the expensive LLMs such as ChatGPT and GPT-4. It offers a drop-in replacement of `openai.Completion` or `openai.ChatCompletion` adding powerful functionalities like tuning, caching, error handling, and templating. For example, you can optimize generations by LLM with your own tuning data, success metrics and budgets. +Autogen also helps maximize the utility out of the expensive LLMs such as ChatGPT and GPT-4. It offers a drop-in replacement of `openai.Completion` or `openai.ChatCompletion` adding powerful functionalities like tuning, caching, error handling, and templating. For example, you can optimize generations by LLM with your own tuning data, success metrics and budgets. ```python # perform tuning config, analysis = autogen.Completion.tune(