@@ -66,20 +66,20 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
66
66
"mount_location" : " /models" ,
67
67
"volume_size_in_gbs" : 500 ,
68
68
"include" : [
69
- " new_example_sglang .yaml" ,
69
+ " offline_sglang_example .yaml" ,
70
70
" NousResearch/Meta-Llama-3.1-8B"
71
71
]
72
72
}
73
73
],
74
74
"output_object_storage" : [
75
75
{
76
76
"bucket_name" : " inference_output" ,
77
- "mount_location" : " /mlcommons_output " ,
77
+ "mount_location" : " /benchmarking_output " ,
78
78
"volume_size_in_gbs" : 200
79
79
}
80
80
],
81
81
"recipe_container_command_args" : [
82
- " /models/new_example_sglang .yaml"
82
+ " /models/offline_sglang_example .yaml"
83
83
],
84
84
"recipe_replica_count" : 1 ,
85
85
"recipe_container_port" : " 8000" ,
@@ -93,7 +93,7 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
93
93
```
94
94
95
95
---
96
- ### [ Sample Blueprint (Job Mode for Offline vLLM Inference)] ( dhttps://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_deployment_sglang .json )
96
+ ### [ Sample Blueprint (Job Mode for Offline vLLM Inference)] ( dhttps://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_deployment_vllm .json )
97
97
98
98
``` json
99
99
{
@@ -116,7 +116,7 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
116
116
"output_object_storage" : [
117
117
{
118
118
"bucket_name" : " inference_output" ,
119
- "mount_location" : " /mlcommons_output " ,
119
+ "mount_location" : " /benchmarking_output " ,
120
120
"volume_size_in_gbs" : 200
121
121
}
122
122
],
@@ -161,7 +161,7 @@ experiment_name: "sglang-bench-doc-test-new"
161
161
run_name : " llama3-8b-sglang-test"
162
162
163
163
164
- save_metrics_path : /mlcommons_output /benchmark_output_llama3_sglang.json
164
+ save_metrics_path : /benchmarking_output /benchmark_output_llama3_sglang.json
165
165
166
166
` ` `
167
167
## [Sample Config File - 2 vLLM (` offline_vllm_example.yaml`)](https://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_vllm_example.yaml)
@@ -194,7 +194,7 @@ distributed_executor_backend: mp
194
194
mlflow_uri: http://mlflow-benchmarking.corrino-oci.com:5000
195
195
experiment_name: test-bm-suite-doc
196
196
run_name: llama3-vllm-test
197
- save_metrics_path: /mlcommons_output /benchmark_output_llama3_vllm.json
197
+ save_metrics_path: /benchmarking_output /benchmark_output_llama3_vllm.json
198
198
199
199
` ` `
200
200
0 commit comments