-
Notifications
You must be signed in to change notification settings - Fork 36
Open
3 / 63 of 6 issues completedDescription
Here is the development roadmap for 2025 Q4. Contributions and feedback are welcome(Open a new discussion)
Focus
- More model support
- Improve Language models performance
- Reinforcement learning training framework integration.
Model coverage
Language models
Speculative decoding
- Reference-based speculative decoding support
RL framework integration
- sglang-jax serves as the inference engine backend for tunix
- Accuracy check: Test math-500 and aime24 for deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
- Support return_logprobs: return logprobs as float32
- Support pathways mode: single host & multi host
- Interruptible Sampling [Feature-RL] Support Interruptible Sampling For Tunix From Rollout Angle #447
Performance improvements
- Continue to optimize the implemented model and use the GPU version as the baseline
LoRA Support
- Support multi LoRA servering
Issue: [Feature] Add Multi LoRA Support #311
Deterministic inference
SiqiLi-Fighting, zhuangweiji, JamesBrianD and starlitxiling
Sub-issues
Metadata
Metadata
Assignees
Labels
No labels