You must log in or # to comment.
I’ve done some testing with the two large models and my initial impression is that they seem very similar in quality to Qwen3.5 35B and 27B. Some notable exceptions:
- llama.cpp has speculative decode support on day 1 and it speeds performance noticeably.
- Day 1 base model release will undoubtedly lead to faster finetunes
Can’t wait for the inevitable Claude/Gemini distils.
My verdict is that even though these models benchmark slightly lower then their Qwen equivalents, their performance and support will likely drive me to pick them.
Are there open Claude models? I thought they were proprietary only
Not exactly. Some people have used Claude’s thinking patterns to train other open models.
This one has been in the top 5 on huggingface for weeks https://huggingface.co/Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled



