Skip to content

Commit 98eb545

Browse files
Merge branch 'main' into ruslan/openapi-evals-type-fix
2 parents ff74458 + 51f784b commit 98eb545

File tree

2 files changed

+81
-41
lines changed

2 files changed

+81
-41
lines changed

.github/workflows/sync-openapi-spec-to-docs.yml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ on:
77
push:
88
branches: [main]
99
paths:
10-
- 'openapi.yaml'
10+
- "openapi.yaml"
1111

1212
jobs:
1313
sync:
@@ -23,7 +23,7 @@ jobs:
2323
echo "${{ secrets.DOCS_DEPLOY_KEY }}" > ~/.ssh/id_ed25519
2424
chmod 600 ~/.ssh/id_ed25519
2525
ssh-keyscan github.com >> ~/.ssh/known_hosts
26-
26+
2727
- name: Install GitHub CLI
2828
run: sudo apt-get install -y gh
2929

@@ -34,7 +34,7 @@ jobs:
3434
3535
- name: Commit and push update
3636
env:
37-
GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}
37+
GH_TOKEN: ${{ secrets.DOCS_REPO_PAT }}
3838
run: |
3939
cd mintlify-docs
4040
git config user.name "github-actions[bot]"
@@ -56,4 +56,5 @@ jobs:
5656
--title "Sync OpenAPI spec" \
5757
--body "This PR syncs openapi.yaml from openapi repo." \
5858
--base main \
59-
--head "$BRANCH_NAME"
59+
--head "$BRANCH_NAME"
60+

openapi.yaml

Lines changed: 76 additions & 37 deletions
Original file line numberDiff line numberDiff line change
@@ -159,8 +159,9 @@ paths:
159159
)
160160
161161
response = client.completions.create(
162-
model="meta-llama/Llama-2-70b-hf",
162+
model="meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
163163
prompt="The largest city in France is",
164+
max_tokens=1
164165
)
165166
166167
print(response.choices[0].text)
@@ -174,8 +175,9 @@ paths:
174175
});
175176
176177
const response = await client.completions.create({
177-
model: "meta-llama/Llama-2-70b-hf",
178+
model: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
178179
prompt: "The largest city in France is",
180+
max_tokens: 1
179181
});
180182
181183
console.log(response.choices[0].text);
@@ -189,8 +191,9 @@ paths:
189191
});
190192
191193
const response = await client.completions.create({
192-
model: "meta-llama/Llama-2-70b-hf",
194+
model: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
193195
prompt: "The largest city in France is",
196+
max_tokens: 1
194197
});
195198
196199
console.log(response.choices[0].text);
@@ -201,8 +204,9 @@ paths:
201204
-H "Authorization: Bearer $TOGETHER_API_KEY" \
202205
-H "Content-Type: application/json" \
203206
-d '{
204-
"model": "meta-llama/Llama-2-70b-hf",
205-
"prompt": "The largest city in France is"
207+
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
208+
"prompt": "The largest city in France is",
209+
"max_tokens": 1
206210
}'
207211
operationId: completions
208212
requestBody:
@@ -1236,6 +1240,12 @@ paths:
12361240
from_checkpoint:
12371241
type: string
12381242
description: The checkpoint identifier to continue training from a previous fine-tuning job. Format is `{$JOB_ID}` or `{$OUTPUT_MODEL_NAME}` or `{$JOB_ID}:{$STEP}` or `{$OUTPUT_MODEL_NAME}:{$STEP}`. The step value is optional; without it, the final checkpoint will be used.
1243+
from_hf_model:
1244+
type: string
1245+
description: The Hugging Face Hub repo to start training from. Should be as close as possible to the base model (specified by the `model` argument) in terms of architecture and size.
1246+
hf_model_revision:
1247+
type: string
1248+
description: The revision of the Hugging Face Hub model to continue training from. E.g., hf_model_revision=main (default, used if the argument is not provided) or hf_model_revision='607a30d783dfa663caf39e06633721c8d4cfcd7e' (specific commit).
12391249
hf_api_token:
12401250
type: string
12411251
description: The API token for the Hugging Face Hub.
@@ -3213,41 +3223,41 @@ paths:
32133223
type: classify
32143224
parameters:
32153225
judge:
3216-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3217-
system_template: "You are a helpful assistant that classifies text."
3218-
labels: ["Toxic", "Non-Toxic"]
3219-
pass_labels: ["Non-Toxic"]
3220-
model_to_evaluate: "output_column"
3221-
input_data_file_path: "file-abcd-1234"
3226+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3227+
system_template: 'You are a helpful assistant that classifies text.'
3228+
labels: ['Toxic', 'Non-Toxic']
3229+
pass_labels: ['Non-Toxic']
3230+
model_to_evaluate: 'output_column'
3231+
input_data_file_path: 'file-abcd-1234'
32223232
score:
32233233
summary: Score evaluation example
32243234
value:
32253235
type: score
32263236
parameters:
32273237
judge:
3228-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3229-
system_template: "You are a helpful assistant that scores responses."
3238+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3239+
system_template: 'You are a helpful assistant that scores responses.'
32303240
min_score: 0
32313241
max_score: 10
32323242
pass_threshold: 7
32333243
model_to_evaluate:
3234-
model_name: "meta-llama/Llama-3-8B-Instruct-Turbo"
3244+
model_name: 'meta-llama/Llama-3-8B-Instruct-Turbo'
32353245
max_tokens: 512
32363246
temperature: 0.7
3237-
system_template: "You are a helpful assistant."
3238-
input_template: "Classify: {prompt}"
3239-
input_data_file_path: "file-1234-abcd"
3247+
system_template: 'You are a helpful assistant.'
3248+
input_template: 'Classify: {prompt}'
3249+
input_data_file_path: 'file-1234-abcd'
32403250
compare:
32413251
summary: Compare evaluation example
32423252
value:
32433253
type: compare
32443254
parameters:
32453255
judge:
3246-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3247-
system_template: "You are a helpful assistant that compares responses."
3248-
model_a: "response_a"
3249-
model_b: "response_b"
3250-
input_data_file_path: "file-1234-abcd"
3256+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3257+
system_template: 'You are a helpful assistant that compares responses.'
3258+
model_a: 'response_a'
3259+
model_b: 'response_b'
3260+
input_data_file_path: 'file-1234-abcd'
32513261
responses:
32523262
'200':
32533263
description: Successful response
@@ -3367,7 +3377,8 @@ paths:
33673377
properties:
33683378
status:
33693379
type: string
3370-
enum: [pending, queued, running, completed, error, user_error]
3380+
enum:
3381+
[pending, queued, running, completed, error, user_error]
33713382
example: completed
33723383
results:
33733384
oneOf:
@@ -3416,7 +3427,15 @@ paths:
34163427
properties:
34173428
status:
34183429
type: string
3419-
enum: [completed, error, running, queued, user_error, inference_error]
3430+
enum:
3431+
[
3432+
completed,
3433+
error,
3434+
running,
3435+
queued,
3436+
user_error,
3437+
inference_error,
3438+
]
34203439
description: The new status for the job
34213440
results:
34223441
oneOf:
@@ -4132,6 +4151,11 @@ components:
41324151
type: string
41334152
description: The name of the moderation model used to validate tokens. Choose from the available moderation models found [here](https://docs.together.ai/docs/inference-models#moderation-models).
41344153
example: 'safety_model_name'
4154+
reasoning_effort:
4155+
type: string
4156+
enum: ['low', 'medium', 'high']
4157+
description: Controls the level of reasoning effort the model should apply when generating responses. Higher values may result in more thoughtful and detailed responses but may take longer to generate.
4158+
example: 'medium'
41354159

41364160
ChatCompletionMessageParam:
41374161
oneOf:
@@ -4513,12 +4537,8 @@ components:
45134537
properties:
45144538
file:
45154539
oneOf:
4516-
- type: string
4517-
format: binary
4518-
description: Audio file to transcribe
4519-
- type: string
4520-
format: uri
4521-
description: Public HTTP/HTTPS URL to audio file
4540+
- $ref: '#/components/schemas/AudioFileBinary'
4541+
- $ref: '#/components/schemas/AudioFileUrl'
45224542
description: Audio file upload or public HTTP/HTTPS URL. Supported formats .wav, .mp3, .m4a, .webm, .flac.
45234543
model:
45244544
type: string
@@ -4565,7 +4585,7 @@ components:
45654585
maxItems: 2
45664586
description: Controls level of timestamp detail in verbose_json. Only used when response_format is verbose_json. Can be a single granularity or an array to get multiple levels.
45674587
default: segment
4568-
example: ["word", "segment"]
4588+
example: ['word', 'segment']
45694589

45704590
AudioTranscriptionResponse:
45714591
oneOf:
@@ -4730,7 +4750,7 @@ components:
47304750
maxItems: 2
47314751
description: Controls level of timestamp detail in verbose_json. Only used when response_format is verbose_json. Can be a single granularity or an array to get multiple levels.
47324752
default: segment
4733-
example: ["word", "segment"]
4753+
example: ['word', 'segment']
47344754

47354755
AudioTranslationResponse:
47364756
oneOf:
@@ -5413,6 +5433,10 @@ components:
54135433
type: string
54145434
from_checkpoint:
54155435
type: string
5436+
from_hf_model:
5437+
type: string
5438+
hf_model_revision:
5439+
type: string
54165440

54175441
FinetuneResponseTruncated:
54185442
type: object
@@ -5533,6 +5557,12 @@ components:
55335557
from_checkpoint:
55345558
type: string
55355559
description: Checkpoint used to continue training
5560+
from_hf_model:
5561+
type: string
5562+
description: Hugging Face Hub repo to start training from
5563+
hf_model_revision:
5564+
type: string
5565+
description: The revision of the Hugging Face Hub model to continue training from
55365566

55375567
FinetuneJobStatus:
55385568
type: string
@@ -6611,14 +6641,14 @@ components:
66116641
type: string
66126642
minItems: 2
66136643
description: List of possible classification labels
6614-
example: ["yes", "no"]
6644+
example: ['yes', 'no']
66156645
pass_labels:
66166646
type: array
66176647
items:
66186648
type: string
66196649
minItems: 1
66206650
description: List of labels that are considered passing
6621-
example: ["yes"]
6651+
example: ['yes']
66226652
model_to_evaluate:
66236653
$ref: '#/components/schemas/EvaluationModelOrString'
66246654
input_data_file_path:
@@ -6736,7 +6766,7 @@ components:
67366766
workflow_id:
67376767
type: string
67386768
description: The ID of the created evaluation job
6739-
example: "eval-1234-1244513"
6769+
example: 'eval-1234-1244513'
67406770
status:
67416771
type: string
67426772
enum: [pending]
@@ -6846,7 +6876,6 @@ components:
68466876
description: JSON string representing label counts
68476877
example: '{"yes": 10, "no": 0}'
68486878

6849-
68506879
EvaluationScoreResults:
68516880
type: object
68526881
properties:
@@ -6916,4 +6945,14 @@ components:
69166945
example: 0
69176946
result_file_id:
69186947
type: string
6919-
description: Data File ID
6948+
description: Data File ID
6949+
6950+
AudioFileBinary:
6951+
type: string
6952+
format: binary
6953+
description: Audio file to transcribe
6954+
6955+
AudioFileUrl:
6956+
type: string
6957+
format: uri
6958+
description: Public HTTPS URL to audio file

0 commit comments

Comments
 (0)