Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This isn’t accurate - most of the style comes from the fine tuning and reinforcement learning, not from the original training data.

At some point people got this idea that LLMs just repeat or imitate their training data, and that’s completely false for today’s models.



>This isn’t accurate - most of the style comes from the fine tuning and reinforcement learning, not from the original training data.

Fine tuning, reinforcement, etc are all 'training' in my books. Perhaps this is your confusion over 'people got this idea'


> Fine tuning, reinforcement, etc are all 'training' in my books.

They are but they have nothing to do with how frequent anything is in literature which was your main point.


Agreed. The pre-2025 base models don't write like this.


So LLMs have gotten creativity recently?


No, my point has nothing to do with creativity. It's about the fact that their output is taylored to look and sound in a certain way in the later stages of model training, it's not representative of the original text data the base model was trained on.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: