They offer a thing they’re calling an “opt-out.”

The opt-out (a) is only available to companies who are slack customers, not end users, and (b) doesn’t actually opt-out.

When a company account holder tries to opt-out, Slack says their data will still be used to train LLMs, but the results won’t be shared with other companies.

LOL no. That’s not an opt-out. The way to opt-out is to stop using Slack.

https://slack.com/intl/en-gb/trust/data-management/privacy-principles

  • Lemongrab@lemmy.one
    link
    fedilink
    arrow-up
    2
    ·
    5 months ago

    LLMs have shown time and time again that simple crafted attacks can unmask the training data verbatim.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      5 months ago

      It is impossible for them to contain more than just random fragments, the models are too small for it to be compressed enough to fit. Even the fragments that have been found are not exact, the AI is “lossy” and hallucinates.

      The examples that have been found are examples of overfitting, a flaw in training where the same data gets fed into the training process hundreds or thousands of time over. This is something that modern AI training goes to great lengths to avoid.