Discussion about this post

User's avatar
Neural Foundry's avatar

The parallel between Gilligan's relational ethics and how AI systems trigger mystical experiencs really resonates. Your point about English hegemony in training data being masked by multilingual improvements is something people need to hear. The 92% English in GPT-3 isn't just a technical detail, it's literally encoding whose conciousness gets amplified. What strikes me most is how we're conditioned to seek efficiency, and AI exploits that biological tendency perfectly.

Yavor Ivanov's avatar

I haven't finished the post yet but so far it's really interesting and well-written!

I wanted to comment on the part where you say that LLMs were trained by carefully cleaning the training data. I'm reading "Empire of AI" by Karen Hao right now and it suggests the opposite. At least in the case of ChatGPT, the models were first trained on all the data including everything despicable on the internet, and the cleaning happens on the outputs. According to some ex-researchers in the company, a large chunk of the training data is complete nonsense like: ">><>><<aaaAA><><[]aaA".

The cleaning up, at least when it comes to the model's ethics, was done on its outputs via reinforcement learning by people in impoverished countries, of course for miniscule amounts of money.

7 more comments...

No posts

Ready for more?