Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What I meant was... humans don't deliberately post WRONG output from chatGPT on the internet. IF they use it to write some blog post or something they will curate the output from chatGPT such that the output fits the topic and is correct to the context. Then when that data gets scraped for training it will be "curated" so bad data generated by the LLM isn't visible.

This is the scenario that occurs when the majority of text on the internet becomes generated by an LLM. Training data from humans is STILL fed back into the LLM via curation of the LLMs own data.

Also please don't ask if I'm "ok" just respond to the comment.



Let me introduce you to the concept of Informational Warfare.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: