• rumba@lemmy.zip
    link
    fedilink
    English
    arrow-up
    16
    ·
    22 hours ago

    Microsoft Probing If DeepSeek-Linked Group Improperly Obtained data the same way OpenAI did. –FTFY

  • rumba@lemmy.zip
    link
    fedilink
    English
    arrow-up
    9
    ·
    22 hours ago

    OMG Competition!

    QUICK, they’re a foreign threat! They’re coming right for us!!!

  • NutWrench@lemmy.world
    link
    fedilink
    English
    arrow-up
    19
    arrow-down
    1
    ·
    1 day ago

    When a writer copies someone else’s work without cites or compensation, it’s called “plagiarism.” But when an AI does it, it’s called “LLM training.”

  • Deepus@lemm.ee
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    1 day ago

    What the fuck is Microsoft getting involved for?! Maybe concentrate on not providing shitty fucking software fuck heads!

    • TransplantedSconie@lemm.ee
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      1
      ·
      2 days ago

      Chinese company:

      Truly, you have a dizzling intellect.

      Microsoft:

      AND IM JUST GETTING STARTED! Where was I?

      Chinese company:

      Stealing data…

  • aeronmelon@lemmy.world
    link
    fedilink
    English
    arrow-up
    62
    arrow-down
    3
    ·
    2 days ago

    “You can’t steal that public data! We stole it first!”

    And considering that’s exactly what Microsoft did to Apple with point and click, what irony!

    • latenightnoir@lemmy.world
      link
      fedilink
      English
      arrow-up
      116
      arrow-down
      1
      ·
      2 days ago

      Yep, NOW it’s a problem, though! Because it’s someone else doing the same thing, someone who isn’t part of the human centipede starting at Trump’s colon.

  • daddy32@lemmy.world
    link
    fedilink
    English
    arrow-up
    24
    ·
    2 days ago

    Lol its like fucking lavrov from fucking russia screaming “this is against international law” when Europe froze their assets.

    • sunzu2@thebrainbin.org
      link
      fedilink
      arrow-up
      15
      ·
      2 days ago

      Bro… US reaction here is so pathetic…

      The behavior is indicative of a bigger issue. They really do think only they are allowed to cheat and steal to win lol

  • Shortstack@reddthat.com
    link
    fedilink
    English
    arrow-up
    27
    ·
    2 days ago

    What’s the game plan if they did?

    Trade restrictions?

    China already proved those did fuck all to stop them from developing their own model.

    Ducking knew this ai bubble would burst sooner or later, just glad we can finally get on with it now.

    • ɔiƚoxɘup@infosec.pub
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 day ago

      I ducking knew it too, I’ve been a long for the ride though. The models still do have some niche applications where they’re actually useful.

      This whole thing with OpenAI and Microsoft whinging about fair play is truly laughable though. What clowns.

      As a side note, it took a few tries to write ducking, my keyboard kept correcting it to fucking. We’re definitely 2 different people. Lol.

    • rumba@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      22 hours ago
      1. Name the company involved a military asset
      2. Forbid US companies from hosting the models
      3. Pressure foreign companies to not work with them
  • ubergeek@lemmy.today
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    2
    ·
    2 days ago

    Are they worried that deepsink too stuff written by others, mixed it up, and repackaged it as it’s own?

    Well, yeah, that’s all AI is. An expensive weighted pachinko machine, that uses human made content, and remixes it.

    • chiliedogg@lemmy.world
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      1
      ·
      edit-2
      2 days ago

      The question isn’t whether they’ve used the same information. It’s whether they’ve faked the process to achieve that 20x efficiency.

      Look at it like a dictionary. Writing one from scratch is a huge task, no matter how many other books exist. How do you even go about finding all of the words?

      But if other people have already written dictionaries, you can just use their word lists and go from there.

      It’s more efficient, but only because it’s a completely different task.

      • ubergeek@lemmy.today
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        2 days ago

        No AI company has ever made any of their own content to train their models, they took what others created, remixed it, and presented it as something new.

        This AI model did the same thing.

        AI lost its job to AI.

        • chiliedogg@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          2 days ago

          Yes, but that doesn’t mean it is more efficient, which is what the whole thing is about.

          Let’s pretend we’re not talking about AI, but tuna fishing. OpenTuna is sending hundreds of ships to the ocean to go fishing. It’s extremely expensive, but it gets results.

          If another fish distributor shows up out of nowhere selling tuna for 1/10 the price, it would be amazing. But if you found out that they could sell them cheap because they were stealing the fish from OpenTuna warehouses, you wouldn’t argue that the secret to catching fish going forward is theft and stop building boats.

            • chiliedogg@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              1 day ago

              So what happens when OpenTuna runs out of fish to steal and there are no more boats?

              Information doesn’t stop being created. AI models need to be constantly trained and updated with new information. One of the biggest issues with GPT3 was the 2021 knowledge cutoff.

              Let’s pretend you’re building a legal analysis AI tool that scrapes the web for information on local, state, and federal law in the US. If your model was from January 2008 and was never updated, then gay marriage wouldn’t be legal in the US, the ACA wouldn’t exist, Super PACs would be illegal, the Consumer Financial Protection Bureau wouldn’t exist, zoning ordinances in pretty much every city would be out of date, and openly carrying a handgun in Texas would get you jailtime.

              It would essentially be a useless tool, and copying that old training data wouldn’t make a better product no matter how cheap it was to do.

              • ubergeek@lemmy.today
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 day ago

                Once tuna runs out, and we run out of boats?

                Maybe we then stop destroying the tuna population?

                Or, to bring this back to point: the environment will be better off once the AI bubble collapses.