• BaroqueInMind
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        2 months ago

        What are you talking about? It follows the Llama 3 Meta license which is pretty fucking open, and essentially every LLM that isn’t a dogshit copyright-stealing Alibaba Quen model uses it.

        Edit: Mistral has an almost similar license that Meta released Llama 3 with.

        Both Llama 3 and Mistral AI’s non-production licenses restrict commercial use and emphasize ethical responsibility, Llama 3’s license has more explicit prohibitions and control over specific applications. Mistral’s non-production license focuses more on research and testing, with fewer detailed restrictions on ethical matters. Both licenses, however, require separate agreements for commercial usage.

        Tl:Dr Mistral doesn’t give two fucks about ethics and needs money more than Meta

        • Possibly linux@lemmy.zip
          link
          fedilink
          English
          arrow-up
          8
          ·
          2 months ago

          Mistral is licensed under the Apache license version 2.0. This license is recognized under the GNU project and under the Open source initiative. This is because it protects your freedom.

          Meanwhile the Meta license places restrictions on use and arbitrary requirements. It is those requirements that lead me to choose not to use it. The issue with LLM licensing is still open but I certainly do not want a EULA style license with rules and restrictions.

          • BaroqueInMind
            link
            fedilink
            arrow-up
            3
            ·
            2 months ago

            You are correct. I checked HuggingFace just now and see they are all released under Apache license. Thank you for the correction.

      • BaroqueInMind
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        2 months ago

        Hermes3 is based on the latest Llama3.1, Mixtral 8x7B is based on Llama 2 released a while ago. Take a guess which one is better. Read the technical paper, it’s only 12 fucking pages.