Biases are stored in large language models from their too narrow input data and lack of global open permanent memory
Biases are stored in large language models from their too narrow input data and lack of global open permanent memory
This is most likely from the strong social biases that are embedded in the training data, which mostly comes from open free sources. Just as “human languages” can be learned at a one shot level (almost unconscious) and intuitive, so are many subtle conventions and choices. I think of it as a kinds of spectrum where the details at the micro level are important, but also at the milli and centi and Kilo and Mega and Giga levels.