• mindbleach
    link
    fedilink
    arrow-up
    1
    ·
    1 month ago

    A local model. No contest. A desktop computer costs vanishingly little, compared to human labor.

    The issue you need to worry about is “good enough.”

    • merc
      link
      fedilink
      arrow-up
      1
      ·
      1 month ago

      Are you assuming you’re just getting access to the model for free? So far some of these things have been available for free, but I think that’s only because the AI companies are desperately trying to get traction before the bubble pops.

      • mindbleach
        link
        fedilink
        arrow-up
        1
        ·
        1 month ago

        When the bubble pops, do you think these giants stand a chance?

        This all started on consumer hardware. It’s driven by aggressively publishing whitepapers. There’s already a decentralized ecosystem of randos fine-tuning released models. And all signs point to faster development and refined data beating raw scale.

        If complexity goes down another order of magnitude from DeepSeek R1, you’ll see FOSS organizations roll their own models from scratch.

        Another, and it’ll be individual hobbyists.

        Whatever’s available for free is the first choice for companies evaluating “good enough,” because it’s hard to get cheaper than free.

        • merc
          link
          fedilink
          arrow-up
          1
          ·
          1 month ago

          It started on consumer hardware, but the models that are supposedly good enough to replace an employee are the ones that took billions to develop.

          Can you get something useful on consumer hardware? Probably. Is it world changing, enough to cause developers to lose their jobs? Maybe? But, it seems unlikely to me.

          • mindbleach
            link
            fedilink
            arrow-up
            1
            ·
            1 month ago

            OpenAI spent billions. DeekSeek spent six million. A lot of whitepapers are about making that number shrink.

            Model size is not a measure of power, either. Current desktop stuff beats last year’s big boys. Small models train faster.

            Whether that’s ever enough to let any dipshit program professionally remains to be seen. But I didn’t think we’d get this far.

            • merc
              link
              fedilink
              arrow-up
              1
              ·
              1 month ago

              DeepSeek claims they spent six million, are they actually accounting for everything they spent, or are they trying to make it seem like they spent less than they really did?

              • mindbleach
                link
                fedilink
                arrow-up
                1
                ·
                1 month ago

                Oh no, what if they only saved two orders of magnitude, instead of three?

                Google knew smaller was better a decade ago. They did AlphaGo with all the data they could find, and a supercomputer famously beat Lee Sedol. A year later they did AlphaGo Zero, one-tenth the size, trained only by playing, and it reliably beat AlphaGo. A year later, they did AlphaZero, one-tenth that size, and it played every board game at a higher level than leading engines. A year later, MuZero played Atari 2600 games, just by looking at the screen.

                OpenAI would not be shitting their pants and trying to make DeepSeek R1 illegal if they still thought bigger meant better.