• A_A@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    edit-2
    10 months ago

    souce has this title now :

    Mistral CEO confirms ‘leak’ of new open source AI model nearing GPT-4 performance


    2 excerpts :

    Mistral co-founder and CEO Arthur Mensch took to X to clarify: “An over-enthusiastic employee of one of our early access customers leaked a quantised (and watermarked) version of an old model we trained and distributed quite openly…
    To quickly start working with a few selected customers, we retrained this model from Llama 2 the minute we got access to our entire cluster — the pretraining finished on the day of Mistral 7B release. We’ve made good progress since — stay tuned!“

    Quantization in ML ((machine learning)) refers to a technique used to make it possible to run certain AI models on less powerful computers and chips by replacing specific long numeric sequences in a model’s architecture with shorter ones.

    • GBU_28@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      10 months ago

      Llama.cpp based on hardware.

      DL from huggingface and run a command