diff --git a/openapi.yaml b/openapi.yaml
index 13d9f86..6fa2db0 100644
--- a/openapi.yaml
+++ b/openapi.yaml
@@ -5643,7 +5643,8 @@ paths:
description: |
Filter hardware configurations by model compatibility. When provided,
the response includes availability status for each compatible configuration.
- example: meta-llama/Llama-3-70b-chat-hf
+ [See all of Together AI's dedicated models](https://docs.together.ai/docs/dedicated-models)
+ example: deepseek-ai/DeepSeek-R1
responses:
'200':
description: 'List of available hardware configurations'
@@ -9097,20 +9098,11 @@ components:
items:
$ref: '#/components/schemas/ChatCompletionMessageParam'
model:
+ type: string
description: >
The name of the model to query.
[See all of Together AI's chat models](https://docs.together.ai/docs/serverless-models#chat-models)
- example: Qwen/Qwen3.5-9B
- anyOf:
- - type: string
- enum:
- - Qwen/Qwen2.5-72B-Instruct-Turbo
- - Qwen/Qwen2.5-7B-Instruct-Turbo
- - Qwen/Qwen3.5-9B
- - meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
- - meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo
- - type: string
max_tokens:
type: integer
description: The maximum number of tokens to generate.
@@ -11313,18 +11305,15 @@ components:
display_name:
type: string
description: A human-readable name for the endpoint
- examples:
- - My Llama3 70b endpoint
+ example: My Llama3 70b endpoint
model:
type: string
description: The model to deploy on this endpoint
- examples:
- - meta-llama/Llama-3-8b-chat-hf
+ example: deepseek-ai/DeepSeek-R1
hardware:
type: string
description: The hardware configuration to use for this endpoint
- examples:
- - 1x_nvidia_a100_80gb_sxm
+ example: 1x_nvidia_a100_80gb_sxm
autoscaling:
$ref: '#/components/schemas/Autoscaling'
description: Configuration for automatic scaling of the endpoint
@@ -11380,19 +11369,19 @@ components:
name:
type: string
description: System name for the endpoint
- example: devuser/meta-llama/Llama-3-8b-chat-hf-a32b82a1
+ example: devuser/deepseek-ai/DeepSeek-R1-a32b82a1
display_name:
type: string
description: Human-readable name for the endpoint
- example: My Llama3 70b endpoint
+ example: My DeepSeek R1 endpoint
model:
type: string
description: The model deployed on this endpoint
- example: meta-llama/Llama-3-8b-chat-hf
+ example: deepseek-ai/DeepSeek-R1
hardware:
type: string
description: The hardware configuration used for this endpoint
- example: 1x_nvidia_a100_80gb_sxm
+ example: 8x_nvidia_h200_140gb_sxm
type:
type: string
enum:
@@ -12069,7 +12058,7 @@ components:
model:
type: string
description: Name of the judge model
- example: 'meta-llama/Llama-3-70B-Instruct-Turbo'
+ example: 'Qwen/Qwen3.5-9B'
system_template:
type: string
description: System prompt template for the judge
@@ -12104,7 +12093,7 @@ components:
model:
type: string
description: Name of the model to evaluate
- example: 'meta-llama/Llama-3-70B-Instruct-Turbo'
+ example: 'Qwen/Qwen3.5-9B'
max_tokens:
type: integer
minimum: 1