Office space meme:

“If y’all could stop calling an LLM “open source” just because they published the weights… that would be great.”

  • Ajen
    link
    fedilink
    arrow-up
    7
    ·
    edit-2
    2 days ago

    I don’t know what, if any, CS background you have, but that is way off. The training dataset is used to generate the weights, or the trained model. In the context of building a trained LLM model, the input is the dataset and the output is the trained model, or weights.

    It’s more appropriate to call deepseek “open-weight” rather than open-source.