• will_a113@lemmy.ml
    link
    fedilink
    English
    arrow-up
    31
    arrow-down
    2
    ·
    edit-2
    23 hours ago

    Gee I wonder if grok shares data with the rest of twitter, Gemini shares with the rest of google or llama… actually I haven’t used Facebook in ages, I don’t even know if there’s a ChatGPT equivalent service on Facebook.

    I do actually wonder if Anthropic shares data with Amazon, or OpenAI with Microsoft (their majority shareholders). That would be a direct 1:1 comparison with what’s happening between deepseek and bytedance (though at least in the latter case you can host-your-own since the model is open source).

    • jacksilver@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      ·
      19 hours ago

      DeepSeek isn’t owned by bytedance, which is why it’s news worthy. The title is stating DeepSeek (an independent company) is sharing data with Bytedance (TikTok parent company).

      So it would be like OpenAI sharing ChatGPT data to Google. Which would be noteworthy as there is likely a lot of sensitive information that people share and most people probably don’t think that data is being sold/shared.

      • will_a113@lemmy.ml
        link
        fedilink
        English
        arrow-up
        3
        ·
        18 hours ago

        Yes, that’s a valid distinction. Though practically speaking, I don’t really know how different it is from anthropic sharing with Amazon, for example

    • Lucidlethargy
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      18 hours ago

      Llama is touted as better than chat gpt. It ostensibly doesn’t share data if you self host it… But I never trust any of them.

      • horse_battery_staple@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        16 hours ago

        Llama is not touted as better than ChatGPT models. On most benchmarks Llama3.3 performs worse than ChatGPT 4o and much worse than OpenAi’s reasoning model o3-mini.

        Unless you’re speaking in terms of privacy and then unfortunately no one really knows as OpenAI does not release any of their metadata and Llama models are self hosted so there’s no method in which the data will leave the LAN they’re hosted on. It’s a case of apples and oranges.

  • horse_battery_staple@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    1
    ·
    edit-2
    23 hours ago

    This is the web chat client/app, just like OpenAI sharing data with Microsoft, or Copilot doing the same. If you self host these LLMs your data stays within your LAN.

    • jwiggler
      link
      fedilink
      English
      arrow-up
      11
      ·
      edit-2
      23 hours ago

      I’m moving to self host all my streaming stuff. Switching from local-only plex to self hosting all my media (spotify, google photos, LLMs) and tools behind a reverse proxy so i can access outside my home. It’s pretty sweet and a good learning experience using reverse proxies

      Edit: Plus fuck these technofeudal lords who enclose access to markets, information, and culture.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      edit-2
      22 hours ago

      You can’t practically self-host Deepseek R1.

      Look, I use the 32B distil on my 3090 every day, but it is not the same thing as full R1. And people need to stop conflating the two.

      And (theoretically) API usage through one of many R1 providers is private.

      • lily33@lemm.ee
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        22 hours ago

        Do you know of a provider is actually private? The few privacy policies I checked all had something like “We might keep some of your data for some time for anti-abuse or other reasons”…

      • jwiggler
        link
        fedilink
        English
        arrow-up
        1
        ·
        20 hours ago

        I dont really use LLMs so I didn’t even realize there were versions with different weights and stuff. I was using 7b, but found it pretty useless. Pretty sure I’m not going to be able run 32B on my rig. lmao.

        guess ill continue being an LLMless pleb.

      • horse_battery_staple@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        20 hours ago

        I use 32b and the 672b side by side. The performance hit is around 20% and I keep all my data local. I am not conflating the two however self hosting works for me just fine. Your usecase is your own certainly. However I’d rather take the performance hit for the added data privacy.

        Also it’s nice to he able to set my own weights and further distil R1

        I have a local python expert a local golang expert and both have my local gitlab repository and I’ve tied their respective Ollama keys to my VSCode IDE.

        • brucethemoose@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          20 hours ago

          Depends for sure. I usually try the 32B first, but give really “hard” queries to some API model.

          • horse_battery_staple@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            19 hours ago

            With the distilled models I have, I’ve been able to build and troubleshoot pretty complicated apps in Golang and Python. However, these distilled models are very specialized and will not do things like write me a story about a duck made out of duct tape or properly summarize articles. There are absolutely limits to my workflow and setup. But I’m pretty happy with it.

        • will_a113@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          18 hours ago

          Have you had any luck importing even a medium-sized codebase and doing reasoning on it? All of my experiments start to show subtle errors past 2k tokens, and at 5k tokens the errors become significant. Any attempt to ingest and process a decent-sized project (say 20k SLOC plus tooling/overhead/config) has been useless, even on models that “should” have a good-enough sized context window.

          • horse_battery_staple@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            16 hours ago

            My codebase is almost 1.2GB of raw python and go files no images. I think it’s somewhere near 15k tokens for the python codebase and 22k for golang due to all the .mod and .io connectors to python libraries… it was a much bigger mess before if you can believe it.

            What size model are you using? I’m getting pretty good results with R1 32b but these have been distilled to be experts in the languages of the codebases. I’m not using any general models for this.

            Also it depends on the language you’re targeting as well. Rust or Lisp have issues due to how much less they’ve been documented. I think golf type languages like brainfuck are impossible. It really comes down to how the language has been documented. Python gave me issues in the beginning until I specified 3.11 in my weights and distillation/training, and that definitely fixed a lot of the hallucinations I was getting from the model.

            I think static typing languages that have consistent documentation would be the easiest for this. Now that I think of it, maybe getting a typescript expert would be something I could tool around with.

            Edited for legibility and the fact that I just went and looked at my datasets again. Much bigger than I initially thought.

  • Lucidlethargy
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    18 hours ago

    This is the worst AI, just in case anyone is wondering. The only thing that’s interesting is that you can see it reasoning and thinking before it answers questions.

    Otherwise, it’s just trash. It makes huge mistakes, it hallucinates often, and it takes more processing power to do those things than other popular alternative AI’s.

    • horse_battery_staple@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      18 hours ago

      All LLMs lie, this is why it’s important to verify what output you’re getting. A GPT is essentially text prediction that has been trained on a very large dataset, think of when you end up sending “ducking autocorrect” in a text. Furthermore, Deepseek has distillations of many models. Which do you have experience using?

      https://github.com/deepseek-ai/DeepSeek-R1

      Edit: To add even more context GPT and Diffusion models are patently not AI as they are not able to verify what output they’re giving. These are all tokens that are fed into a recursive algorithm. They’re vector database queries that have reinforced pathing. None of these “A.I.” models are thinking or reasoning, yet.