Skip to content

Commit ee05525

Browse files
committed
update
Signed-off-by: qingjun <[email protected]>
1 parent c4aa24d commit ee05525

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

_posts/2025-06-26-minimax-m1.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ huggingface-cli download MiniMaxAI/MiniMax-M1-40k
4545

4646
### Deployment
4747

48-
Below is a quick guide to deploying MiniMax-M1 with vLLM and Docker. Each step is annotated for clarity:
48+
Below is a quick guide to deploying MiniMax-M1 with vLLM and Docker:
4949

5050
```bash
5151
# Set environment variables
@@ -131,7 +131,7 @@ As a cutting-edge attention mechanism, Lightning Attention is implemented in vLL
131131

132132
Looking ahead, further optimizations for hybrid architecture support are actively being explored within the vLLM community. Notably, the development of a hybrid allocator is expected to enable even more efficient memory management tailored to the unique requirements of models like MiniMax-M1.
133133

134-
In addition, full support for vLLM v1 is planned, with the hybrid model architecture expected to be migrated into the v1 framework. These advancements are anticipated to unlock further performance improvements and provide a more robust foundation for future developments.
134+
In addition, full support for [vLLM v1](https://minimax-m1.vllm-blog-source.pages.dev/2025/01/27/v1-alpha-release) is planned, with the hybrid model architecture expected to be migrated into the v1 framework. These advancements are anticipated to unlock further performance improvements and provide a more robust foundation for future developments.
135135

136136
## Conclusion
137137

0 commit comments

Comments
 (0)