• saigot@lemmy.ca
    link
    fedilink
    arrow-up
    4
    ·
    1 年前

    If it was done with enough regularity to eb a problem, one could just put an LLM model like this in-between to preprocess the data.

    • Azzu@lemm.ee
      link
      fedilink
      arrow-up
      4
      ·
      1 年前

      That doesn’t work, you can’t train models on another model’s output without degrading the quality. At least not currently.

      • Vashtea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 年前

        I don’t think he was suggesting training on another model’s output, just using ai to filter the training data before it is used.

      • FooBarrington@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 年前

        No, that’s not true. All current models use output from previous models as part of their training data. You can’t solely rely on it, but that’s not strictly necessary.