zlacker

[return to "AI agents are starting to eat SaaS"]
1. Oarch+65[view] [source] 2025-12-15 00:24:57
>>jnord+(OP)
Earlier this year I thought that rare proprietary knowledge and IP was a safe haven from AI, since LLMs can only scrub public data.

Then it dawned on me how many companies are deeply integrating Copilot into their everyday workflows. It's the perfect Trojan Horse.

◧◩
2. Aurorn+S5[view] [source] 2025-12-15 00:31:19
>>Oarch+65
Using an LLM on data does not ingest that data into the training corpus. LLMs don’t “learn” from the information they operate on, contrary to what a lot of people assume.

None of the mainstream paid services ingest operating data into their training sets. You will find a lot of conspiracy theories claiming that companies are saying one thing but secretly stealing your data, of course.

◧◩◪
3. popalc+te[view] [source] 2025-12-15 01:40:02
>>Aurorn+S5
Wrong, buddy.

Many of the top AI services use human feedback to continuously apply "reinforcement learning" after the initial deployment of a pre-trained model.

https://en.wikipedia.org/wiki/Reinforcement_learning_from_hu...

◧◩◪◨
4. Aurorn+zf[view] [source] 2025-12-15 01:48:28
>>popalc+te
RLHF is a training step.

Inference (what happens when you use an LLM as a customer) is separate from training.

Inference and training are separate processes. Using an LLM doesn’t train it. That’s not what RLHF means.

[go to top]