zlacker

[parent] [thread] 6 comments
1. yesco+(OP)[view] [source] 2025-08-27 22:19:33
LLMs write in a very coherent, easy to understand way. I see no reason why someone wouldn't want to copy their style or vocabulary if they want to improve their communication skills.

Despite all the complaints about AI slop, there is something ironic about the fact that simply being exposed to it might be a net positive influence for most of society. Discord often begins from the simplest of communication errors after all...

replies(2): >>capnre+X2 >>mingus+P5
2. capnre+X2[view] [source] 2025-08-27 22:41:54
>>yesco+(OP)
Sure, if you're learning to write and want lots of examples of a particular style, LLMs can generate that for you. Just don't assume that is a normal writing style, or that it matches a particular genre (say, workplace communication, or academic writing, or whatever).

Our experience (https://arxiv.org/abs/2410.16107) is that LLMs like GPT-4o have a particular writing style, including both vocabulary and distinct grammatical features, regardless of the type of text they're prompted with. The style is informationally dense, features longer words, and favors certain grammatical structures (like participles; GPT-4o loooooves participles).

With Llama we're able to compare base and instruction-tuned models, and it's the instruction-tuned models that show the biggest differences. Evidently the AI companies are (deliberately or not) introducing particular writing styles with their instruction-tuning process. I'd like to get access to more base models to compare and figure out why.

replies(2): >>yesco+F5 >>ACCoun+9a
◧◩
3. yesco+F5[view] [source] [discussion] 2025-08-27 23:06:16
>>capnre+X2
I definitely know what you mean, each model definitely has it's own style. I find myself mentally framing them as like horses with different personalities and riding quirks.

Still, perhaps saying "copy" was a bit misleading. Influence would have been more precise way of putting it. After all, there is no such thing as a "normal" writing style in the first place.

So long as you communicate with anything or anyone, I find people will naturally just absorb the parts they like without even noticing most of the time.

4. mingus+P5[view] [source] 2025-08-27 23:07:05
>>yesco+(OP)
When I learned that AI was trained off of internet posts, and then LLMs were the new bots making internet posts, it immediately made me think that the entire internet would degrade like a jpeg that you keep compressing and sending around

I guess this is called model collapse

But now I’m wondering if people are collapsing. LLMs start to sound like us. We adapt and start to sound like LLMs that gets fed into the next set of model training…

What is the dystopian version of this end game?

replies(1): >>yesco+T6
◧◩
5. yesco+T6[view] [source] [discussion] 2025-08-27 23:16:34
>>mingus+P5
Perhaps a surreal one where we drill past the bedrock and start to communicate in raw tokens, conveying extreme levels of depth and nuance within a single sentence?

When humans carved words into stone, the words and symbols were often suited for the medium, a bunch of straight lines assembled together in various patterns. But with the ink, you get circles, and elaborate curved lines, symbols suited to the movement patterns we can make quickly with our wrist.

But what of the digital keyboard? Any symbol that can be drawn in 2 dimensions. They can be typed quickly, with exact precision. Human language was already destined to head in a weird direction.

◧◩
6. ACCoun+9a[view] [source] [discussion] 2025-08-27 23:45:46
>>capnre+X2
Go vibe check Kimi-K2. One of the weirdest models out there now, and it's open weights - with both "base" and "instruct" versions available.

The language it uses is peculiar. It's like the entire model is a little bit ESL.

I suspect that this pattern comes from SFT and RLHF, not the optimizer or the base architecture or the pre-training dataset choices, and the base model itself would perform much more "in line" with other base models. But I could be wrong.

Goes to show just how "entangled" those AIs are, and how easy it is to affect them in unexpected ways with training. Base models have a vast set of "styles" and "language usage patterns" they could draw from - but instruct-tuning makes a certain set of base model features into the "default" persona, shaping the writing style this AI would use down the line.

replies(1): >>jjani+sE
◧◩◪
7. jjani+sE[view] [source] [discussion] 2025-08-28 05:01:03
>>ACCoun+9a
Kimi tends to be very.. casual from my usage, like informal millenial style, without being prompted to do so.
[go to top]