• afk_strats@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    1 day ago

    I’ve done some testing with the two large models and my initial impression is that they seem very similar in quality to Qwen3.5 35B and 27B. Some notable exceptions:

    1. llama.cpp has speculative decode support on day 1 and it speeds performance noticeably.
    2. Day 1 base model release will undoubtedly lead to faster finetunes

    Can’t wait for the inevitable Claude/Gemini distils.

    My verdict is that even though these models benchmark slightly lower then their Qwen equivalents, their performance and support will likely drive me to pick them.