Skip to content

Commit 9e1f230

Browse files
Add Embeddinggemma-300m model to Workers AI docs (#24937)
This PR also adds changes in schemas for partner models
1 parent 3824d28 commit 9e1f230

File tree

4 files changed

+93
-6
lines changed

4 files changed

+93
-6
lines changed
Lines changed: 58 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,58 @@
1+
{
2+
"id": "d2f07a41-c152-4061-8083-ec655cbf91a1",
3+
"source": 1,
4+
"name": "@cf/google/embeddinggemma-300m",
5+
"description": "EmbeddingGemma is a 300M parameter, state-of-the-art for its size, open embedding model from Google, built from Gemma 3 (with T5Gemma initialization) and the same research and technology used to create Gemini models. EmbeddingGemma produces vector representations of text, making it well-suited for search and retrieval tasks, including classification, clustering, and semantic similarity search. This model was trained with data in 100+ spoken languages.",
6+
"task": {
7+
"id": "0137cdcf-162a-4108-94f2-1ca59e8c65ee",
8+
"name": "Text Embeddings",
9+
"description": "Feature extraction models transform raw data into numerical features that can be processed while preserving the information in the original dataset. These models are ideal as part of building vector search applications or Retrieval Augmented Generation workflows with Large Language Models (LLM)."
10+
},
11+
"created_at": "2025-09-03 20:48:38.784",
12+
"tags": [],
13+
"properties": [],
14+
"schema": {
15+
"input": {
16+
"type": "object",
17+
"properties": {
18+
"text": {
19+
"oneOf": [{
20+
"type": "string"
21+
}, {
22+
"type": "array",
23+
"items": {
24+
"type": "string"
25+
}
26+
}],
27+
"description": "Input text to embed. Can be a single string or a list of strings."
28+
}
29+
},
30+
"required": ["text"]
31+
},
32+
"output": {
33+
"type": "object",
34+
"properties": {
35+
"data": {
36+
"type": "array",
37+
"items": {
38+
"type": "array",
39+
"items": {
40+
"type": "number"
41+
}
42+
},
43+
"description": "Embedding vectors, where each vector is a list of floats."
44+
},
45+
"shape": {
46+
"type": "array",
47+
"items": {
48+
"type": "integer"
49+
},
50+
"minItems": 2,
51+
"maxItems": 2,
52+
"description": "Shape of the embedding data as [number_of_embeddings, embedding_dimension]."
53+
}
54+
},
55+
"required": ["data", "shape"]
56+
}
57+
}
58+
}

src/content/workers-ai-models/lucid-origin.json

Lines changed: 10 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -11,10 +11,6 @@
1111
"created_at": "2025-08-25 19:21:28.770",
1212
"tags": [],
1313
"properties": [
14-
{
15-
"property_id": "partner",
16-
"value": "true"
17-
},
1814
{
1915
"property_id": "price",
2016
"value": [
@@ -29,6 +25,10 @@
2925
"currency": "USD"
3026
}
3127
]
28+
},
29+
{
30+
"property_id": "partner",
31+
"value": "true"
3232
}
3333
],
3434
"schema": {
@@ -68,7 +68,12 @@
6868
},
6969
"num_steps": {
7070
"type": "integer",
71-
"default": 4,
71+
"minimum": 1,
72+
"maximum": 40,
73+
"description": "The number of diffusion steps; higher values can improve quality but take longer"
74+
},
75+
"steps": {
76+
"type": "integer",
7277
"minimum": 1,
7378
"maximum": 40,
7479
"description": "The number of diffusion steps; higher values can improve quality but take longer"

src/content/workers-ai-models/nova-3.json

Lines changed: 21 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@
111111
"type": "string",
112112
"description": "Arbitrary key-value pairs that are attached to the API response for usage in downstream processing"
113113
},
114-
"filter_words": {
114+
"filler_words": {
115115
"type": "boolean",
116116
"description": "Filler Words can help transcribe interruptions in your audio, like 'uh' and 'um'"
117117
},
@@ -195,6 +195,26 @@
195195
"utt_split": {
196196
"type": "number",
197197
"description": "Seconds to wait before detecting a pause between words in submitted audio."
198+
},
199+
"channels": {
200+
"type": "number",
201+
"description": "The number of channels in the submitted audio"
202+
},
203+
"interim_results": {
204+
"type": "boolean",
205+
"description": "Specifies whether the streaming endpoint should provide ongoing transcription updates as more audio is received. When set to true, the endpoint sends continuous updates, meaning transcription results may evolve over time. Note: Supported only for webosockets."
206+
},
207+
"endpointing": {
208+
"type": "string",
209+
"description": "Indicates how long model will wait to detect whether a speaker has finished speaking or pauses for a significant period of time. When set to a value, the streaming endpoint immediately finalizes the transcription for the processed time range and returns the transcript with a speech_final parameter set to true. Can also be set to false to disable endpointing"
210+
},
211+
"vad_events": {
212+
"type": "boolean",
213+
"description": "Indicates that speech has started. You'll begin receiving Speech Started messages upon speech starting. Note: Supported only for webosockets."
214+
},
215+
"utterance_end_ms": {
216+
"type": "boolean",
217+
"description": "Indicates how long model will wait to send an UtteranceEnd message after a word has been transcribed. Use with interim_results. Note: Supported only for webosockets."
198218
}
199219
},
200220
"required": [

src/content/workers-ai-models/smart-turn-v2.json

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,10 @@
2424
"currency": "USD"
2525
}
2626
]
27+
},
28+
{
29+
"property_id": "realtime",
30+
"value": "true"
2731
}
2832
],
2933
"schema": {

0 commit comments

Comments
 (0)