They offer a thing they’re calling an “opt-out.”
The opt-out (a) is only available to companies who are slack customers, not end users, and (b) doesn’t actually opt-out.
When a company account holder tries to opt-out, Slack says their data will still be used to train LLMs, but the results won’t be shared with other companies.
LOL no. That’s not an opt-out. The way to opt-out is to stop using Slack.
https://slack.com/intl/en-gb/trust/data-management/privacy-principles
You could say it’s to “circumvent” the law or you could say it’s to comply with the law. As long as the PII is gone what’s the problem?
LLMs have shown time and time again that simple crafted attacks can unmask the training data verbatim.
It is impossible for them to contain more than just random fragments, the models are too small for it to be compressed enough to fit. Even the fragments that have been found are not exact, the AI is “lossy” and hallucinates.
The examples that have been found are examples of overfitting, a flaw in training where the same data gets fed into the training process hundreds or thousands of time over. This is something that modern AI training goes to great lengths to avoid.