Skip to content

Commit

Permalink
changed icons for supported providers table
Browse files Browse the repository at this point in the history
  • Loading branch information
lmilan committed Dec 20, 2024
1 parent 968896b commit 9368119
Show file tree
Hide file tree
Showing 2 changed files with 46 additions and 40 deletions.
72 changes: 36 additions & 36 deletions app/_data/plugins/ai-proxy.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,133 +4,133 @@ providers:
url_pattern: 'https://api.openai.com:443/{route_type_path}'
min_version: '3.6'
chat:
supported:
supported: true
streaming: true
upstream_path: '/v1/chat/completions'
route_type: 'llm/v1/chat'
model_example: 'gpt-4'
completions:
supported:
supported: true
streaming: true
upstream_path: '/v1/completions'
route_type: 'llm/v1/completions'
model_example: 'gpt-3.5-turbo-instruct'
streaming:
supported:
- name: 'Cohere'
url_pattern: 'https://api.cohere.com:443/{route_type_path}'
min_version: '3.6'
chat:
supported:
supported: true
streaming: true
upstream_path: '/v1/chat'
route_type: 'llm/v1/chat'
model_example: 'command'
completions:
supported:
supported: true
streaming: true
upstream_path: '/v1/generate'
route_type: 'llm/v1/completions'
model_example: 'command'
streaming:
supported:
- name: 'Azure'
url_pattern: 'https://{azure_instance}.openai.azure.com:443/openai/deployments/{deployment_name}/{route_type_path}'
min_version: '3.6'
chat:
supported:
supported: true
streaming: true
upstream_path: '/openai/deployments/{deployment_name}/chat/completions'
route_type: 'llm/v1/chat'
model_example: 'gpt-4'
completions:
supported:
supported: true
streaming: true
upstream_path: '/openai/deployments/{deployment_name}/completions'
route_type: 'llm/v1/completions'
model_example: 'gpt-3.5-turbo-instruct'
streaming:
supported:
- name: 'Anthropic'
url_pattern: 'https://api.anthropic.com:443/{route_type_path}'
min_version: '3.6'
chat:
supported:
supported: true
streaming: true
upstream_path: '/v1/complete in version 3.6, /v1/messages since version 3.7'
route_type: 'llm/v1/chat'
model_example: 'claude-2.1'
completions:
supported:
supported: true
streaming: false
upstream_path: '/v1/complete'
route_type: 'llm/v1/completions'
model_example: 'claude-2.1'
streaming:
supported: Chat type only
- name: 'Mistral'
formats: 'mistral.ai, OpenAI, raw, and OLLAMA formats'
url_pattern: 'As defined in <code>$UPSTREAM_URL</code>'
min_version: '3.6'
chat:
supported:
supported: true
streaming: true
upstream_path: 'User-defined'
route_type: 'llm/v1/chat'
model_example: 'User-defined'
completions:
supported:
supported: true
streaming: true
upstream_path: 'User-defined'
route_type: 'llm/v1/completions'
model_example: 'User-defined'
streaming:
supported:
- name: 'Llama2'
formats: 'supports Llama2 and Llama3 models and raw, OLLAMA, and OpenAI formats'
url_pattern: 'As defined in <code>$UPSTREAM_URL</code>'
min_version: ''
chat:
supported:
supported: true
streaming: true
upstream_path: 'User-defined'
route_type: 'llm/v1/chat'
model_example: 'User-defined'
completions:
supported:
supported: true
streaming: true
upstream_path: 'User-defined'
route_type: 'llm/v1/completions'
model_example: 'User-defined'
streaming:
supported:
- name: 'Amazon Bedrock'
url_pattern: 'https://bedrock-runtime.{region}.amazonaws.com'
min_version: '3.8'
chat:
supported:
supported: true
streaming: true
upstream_path: 'Use the LLM <code>chat</code> upstream path'
route_type: 'llm/v1/chat'
model_example: '<a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html">Use the model name for the specific LLM provider</a>'
completions:
supported:
streaming:
supported: Chat type only
supported: false
streaming: false
- name: 'Gemini'
url_pattern: 'https://generativelanguage.googleapis.com'
min_version: '3.8'
chat:
supported:
supported: true
streaming: true
upstream_path: 'llm/v1/chat'
route_type: 'llm/v1/chat'
model_example: 'gemini-1.5-flash or gemini-1.5-pro'
completions:
supported:
streaming:
supported: Chat type only
supported: false
streaming: false
- name: 'Hugging Face'
url_pattern: 'https://api-inference.huggingface.co'
min_version: '3.9'
chat:
supported:
supported: true
streaming: true
upstream_path: '/models/{model_provider}/{model_name}'
route_type: 'llm/v1/chat'
model_example: '<a href="https://huggingface.co/models?inference=warm&pipeline_tag=text-generation&sort=trending">Use the model name for the specific LLM provider</a>'
completions:
supported:
supported: true
streaming: true
upstream_path: '/models/{model_provider}/{model_name}'
route_type: 'llm/v1/completions'
model_example: '<a href="https://huggingface.co/models?inference=warm&pipeline_tag=text-generation&sort=trending">Use the model name for the specific LLM provider</a>'
streaming:
supported:

parameters:
provider: 'config.provider'
Expand Down
14 changes: 10 additions & 4 deletions app/_includes/plugins/ai-proxy/tables/supported-providers.html
Original file line number Diff line number Diff line change
Expand Up @@ -4,21 +4,27 @@
<th>Provider</th>
<th>Chat</th>
<th>Completion</th>
<th>Streaming</th>
<th>Chat streaming</th>
<th>Completions streaming</th>
<th>Minimum {{ site.base_gateway }} version</th>
</tr>
</thead>
<tbody>
{% for provider in include.providers %}
{% assign chat = provider.chat.supported %}
{% assign completions = provider.completions.supported %}
{% assign chat_streaming = provider.chat.streaming %}
{% assign completions_streaming = provider.completions.streaming %}
<tr id="{{ provider.name }}">
<td>{{ provider.name }}
{% if provider.formats %}
({{ provider.formats }})
{% endif %}
</td>
<td>{{ provider.chat.supported }}</td>
<td>{{ provider.completions.supported }}</td>
<td>{{ provider.streaming.supported }}</td>
<td>{% if chat %}<span class="inline-flex text-terciary w-5 h-5">{% include_svg '/assets/icons/check.svg' %}</span>{% else %}<span class="inline-flex text-semantic-red-primary w-5 h-5">{% include_svg '/assets/icons/close.svg' %}</span>{% endif %}</td>
<td>{% if completions %}<span class="inline-flex text-terciary w-5 h-5">{% include_svg '/assets/icons/check.svg' %}</span>{% else %}<span class="inline-flex text-semantic-red-primary w-5 h-5">{% include_svg '/assets/icons/close.svg' %}</span>{% endif %}</td>
<td>{% if chat_streaming %}<span class="inline-flex text-terciary w-5 h-5">{% include_svg '/assets/icons/check.svg' %}</span>{% else %}<span class="inline-flex text-semantic-red-primary w-5 h-5">{% include_svg '/assets/icons/close.svg' %}</span>{% endif %}</td>
<td>{% if completions_streaming %}<span class="inline-flex text-terciary w-5 h-5">{% include_svg '/assets/icons/check.svg' %}</span>{% else %}<span class="inline-flex text-semantic-red-primary w-5 h-5">{% include_svg '/assets/icons/close.svg' %}</span>{% endif %}</td>
<td>
{% if provider.min_version < page.min_version.gateway %}
{{ page.min_version.gateway }}
Expand Down

0 comments on commit 9368119

Please sign in to comment.