• davel@lemmygrad.ml
    link
    fedilink
    English
    arrow-up
    0
    ·
    21 hours ago

    Last month: https://lemmygrad.ml/comment/5895113

    Ask it again in Chinese, wherein the model was presumably trained on a Chinese corpus instead of a Five Eyes one.

    This is exactly the problem. These are just engines for regurgitating whatever they have been fed. If they are fed garbage, then all you get out is garbage. For instance notice the use of the buzzword “authoritarian” implicitly assumed to mean “bad”, because that is how it is used in all liberal discourse. If you want a model that does not reproduce liberalism then ceasing training on english language inputs, which are overwhelmingly infected with liberal ideological assumptions, would be a start. It’s still not going to be ideal because what you would really need is proper curation of training content in which a human filters out the garbage. Showing once again the limitations of this technology, but also the danger if used improperly of falsely presenting the hegemonic ideology as “unbiased” facts, or at best taking a noncommittal “middle ground” stance because it has been fed both facts and bullshit, and is of course unable to distinguish between the two.

    Yup. LLM output only reflects its input, and nearly all of the English language corpus in the world is bourgeois cultural hegemony. Truth has nothing to do with it.