Skip to content

Commit

Permalink
Merge pull request #282 from openai/dev/zhang/fix
Browse files Browse the repository at this point in the history
Fixes documentation error
  • Loading branch information
zhang-openai authored Jun 17, 2024
2 parents 2d3212f + c5cc333 commit 0f5de60
Showing 1 changed file with 18 additions and 12 deletions.
30 changes: 18 additions & 12 deletions openapi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7206,17 +7206,6 @@ components:
If specified, our system will make a best effort to sample deterministically, such that repeated requests with the same `seed` and parameters should return the same result.

Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.
service_level:
description: |
Specifies the latency tier to use for processing the request. This parameter is relevant for customers subscribed to the scale tier service:
- If set to 'auto', the system will utilize scale tier credits until they are exhausted.
- If set to 'default', the request will be processed in the shared cluster.

When this parameter is set, the response body will include the `service_tier` utilized.
type: string
enum: ["auto", "default"]
nullable: true
default: null
stop:
description: &completions_stop_description >
Up to 4 sequences where the API will stop generating further tokens. The returned text will not contain the stop sequence.
Expand Down Expand Up @@ -7936,6 +7925,17 @@ components:
Determinism is not guaranteed, and you should refer to the `system_fingerprint` response parameter to monitor changes in the backend.
x-oaiMeta:
beta: true
service_tier:
description: |
Specifies the latency tier to use for processing the request. This parameter is relevant for customers subscribed to the scale tier service:
- If set to 'auto', the system will utilize scale tier credits until they are exhausted.
- If set to 'default', the request will be processed in the shared cluster.

When this parameter is set, the response body will include the `service_tier` utilized.
type: string
enum: ["auto", "default"]
nullable: true
default: null
stop:
description: |
Up to 4 sequences where the API will stop generating further tokens.
Expand Down Expand Up @@ -8077,7 +8077,7 @@ components:
model:
type: string
description: The model used for the chat completion.
scale_tier:
service_tier:
description: The service tier used for processing the request. This field is only included if the `service_tier` parameter is specified in the request.
type: string
enum: ["scale", "default"]
Expand Down Expand Up @@ -8259,6 +8259,12 @@ components:
model:
type: string
description: The model to generate the completion.
service_tier:
description: The service tier used for processing the request. This field is only included if the `service_tier` parameter is specified in the request.
type: string
enum: ["scale", "default"]
example: "scale"
nullable: true
system_fingerprint:
type: string
description: |
Expand Down

0 comments on commit 0f5de60

Please sign in to comment.