@@ -101,12 +101,11 @@ def prompt(
101
101
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
102
102
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
103
103
(better models yield better results, faster models yield faster results). -
104
- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
105
- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
106
- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
107
- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
108
- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
109
- "nova-micro".
104
+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
105
+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
106
+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
107
+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
108
+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
110
109
111
110
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
112
111
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -245,12 +244,11 @@ def score(
245
244
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
246
245
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
247
246
(better models yield better results, faster models yield faster results). -
248
- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
249
- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
250
- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
251
- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
252
- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
253
- "nova-micro".
247
+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
248
+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
249
+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
250
+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
251
+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
254
252
255
253
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
256
254
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -405,12 +403,11 @@ async def prompt(
405
403
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
406
404
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
407
405
(better models yield better results, faster models yield faster results). -
408
- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
409
- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
410
- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
411
- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
412
- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
413
- "nova-micro".
406
+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
407
+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
408
+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
409
+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
410
+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
414
411
415
412
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
416
413
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -549,12 +546,11 @@ async def score(
549
546
"claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
550
547
"nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
551
548
(better models yield better results, faster models yield faster results). -
552
- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
553
- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
554
- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
555
- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
556
- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
557
- "nova-micro".
549
+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
550
+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
551
+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
552
+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
553
+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
558
554
559
555
log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
560
556
For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
0 commit comments