• btobolaski@threads.ruin.io
    link
    fedilink
    English
    arrow-up
    9
    ·
    1 year ago

    Double the training data, double the trained context (4096 now), a chat tuned varient, the omission of the 35b model for now (it apparently isn’t “safe” enough), and commercial use is allowed (not that most of the people using llama cares about licensing).

    • kakes
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Thanks, that sounds more exciting than just “safety” haha.