Skip to content

Conversation

@optrader8
Copy link

TLDR

Dive Deeper

Reviewer Test Plan

Testing Matrix

🍏 πŸͺŸ 🐧
npm run ❓ ❓ ❓
npx ❓ ❓ ❓
Docker ❓ ❓ ❓
Podman ❓ - -
Seatbelt ❓ - -

Linked issues / bugs

optrader8 and others added 2 commits November 9, 2025 22:13
μƒˆλ‘œμš΄ GPT OSS 20B λͺ¨λΈμ— λŒ€ν•œ λ¬Έμ„œ νŒŒμΌμ„ μΆ”κ°€ν–ˆμŠ΅λ‹ˆλ‹€.

πŸ• 컀밋 μ‹œκ°„: 2025. 11. 09. 22:13:47

πŸ“Š λ³€κ²½ 톡계:
  β€’ 총 파일: 1개

πŸ“ μΆ”κ°€λœ 파일:
  + README.gpt-oss-20b.md

πŸ”§ ν•¨μˆ˜ λ³€κ²½ λ‚΄μš©:
  πŸ“„ README.gpt-oss-20b.md (md파일):
    βœ… Added: Customizations(), Optimizations(), Features()

πŸ”§ μ£Όμš” λ³€κ²½ λ‚΄μš©:
  β€’ 개발 λ¬Έμ„œ 및 κ°€μ΄λ“œ κ°œμ„ 
Integrates GPT-OSS-20B as a supported model while maintaining full
backward compatibility with existing Qwen-Coder models.

Key Changes:
- Added GPT-OSS-20B model constants and 128K token limit configuration
- Implemented reasoning_content parsing for both streaming and non-streaming
- Added comprehensive environment configuration and security guidelines
- Optimized timeout settings for lower-end GPU hardware

Features:
- Parallel model support (GPT-OSS-20B + Qwen-Coder)
- reasoning_content transparency feature support
- Graceful degradation for non-GPT-OSS-20B models
- Zero breaking changes, fully backward compatible

Files Modified:
- packages/core/src/config/models.ts (+2 lines)
- packages/core/src/core/tokenLimits.ts (+1 line)
- packages/core/src/core/openaiContentGenerator/converter.ts (+20 lines)

Documentation:
- docs-optrader/00-SUMMARY.md - Executive summary and quick start
- docs-optrader/01-api-test-results.md - Comprehensive API testing
- docs-optrader/02-phase1-configuration.md - Configuration changes
- docs-optrader/03-phase2-core-changes.md - Core implementation details
- docs-optrader/04-environment-configuration.md - Setup and security guide

Testing:
- API tests: βœ… Complete (chat, tool calling, streaming)
- Build verification: βœ… Success
- Integration tests: ⏳ Pending manual validation

Performance:
- Default timeout increased recommendation: 300s (for slow GPUs)
- Token limit: 128K context window
- Generation speed: ~60 tokens/sec (GPU dependent)

Security:
- No hardcoded API keys
- Environment variable based configuration
- Comprehensive security best practices documented

Ref: README.gpt-oss-20b.md
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants