Conversation
|
Thank you so much for this contribution! Integrating Qwen3.5 into the pipeline is indeed timely and valuable. It's great to hear that the training loss looks normal in the first 500 steps. Could you possibly run the training for a few more steps (e.g., 5k or 10k) and share some initial results (like loss curves or preliminary evaluation on SimplerEnv) in this PR? That would be super helpful for the community to get a quick sense of how Qwen3.5 performs with our framework. Thanks again for your work! |
|
Of course, I'd be happy to do so. I'll share some of the latest results once tomorrow's experiments are complete. |
* feat(vlm): add Qwen3.5 model integration * docs(LangForce): updated several textual descriptions for LangForce
* feat(vlm): add Qwen3.5 model integration * docs(LangForce): updated several textual descriptions for LangForce
|
@LiamLian0727 thank you for your work. I am working on add qwen3.5 with fast tokens. This works well with starVLA/model/modules/vlm/tools/add_qwen_special_tokens/add_special_tokens_to_qwen.py. However, transformers AutoProcessor fails to load physical-intelligence/fast. I have tried transformers==5.2.0, 5.3.0, since qwen3.5 requires >=5.2.0. All show this error message: I have installed sentencepiece and tiktoken and tried: I am guessing this might be a bug of transformers. Have you encountered same error? |

This PR includes two updates:
Just now, the official Qwen team released the weights for Qwen3.5 0.8B, 2B, 4B, and 9B. So, it is time to integrate Qwen3.5 into starVLA. I have tested Qwen3.5-4B with QwenGR00T on SimplerEnv (global batch=128). The training loss behavior is normal during the first 500 steps.
Added/updated several textual descriptions for LangForce to make prompts and related outputs clearer.