• arc@lemmy.world
    link
    fedilink
    arrow-up
    3
    arrow-down
    3
    ·
    3 个月前

    You’d probably need to verify all submissions

    Unless you throw an LLM into the mix lol

    • optional@sh.itjust.works
      link
      fedilink
      arrow-up
      12
      arrow-down
      1
      ·
      3 个月前

      You believe an LLM can be used to distinguish facts from fiction? I wonder up to which year that misconception was taught in school.

      The whole point of LLMs is, to convince their users that the “facts” they generate are actual facts.

      • arc@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        3 个月前

        They can browse the web, and I never meant it would be 100 accurate just easier. Don’t think this is going to be a mission critical website

        • Ragnarok314159@sopuli.xyz
          link
          fedilink
          arrow-up
          3
          ·
          3 个月前

          That just it, these “facts” won’t be on the web for stuff approximately 2005 and before. No where on the web is the racist and homophobic shit I was taught in the 80’s and 90’s listed on some wiki.

          LLM’s are mostly useless anyways at distinguishing real information, they are just shit summary tools and poor search engines.

    • ZILtoid1991@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      3 个月前

      LLMs are not magic, otherwise one just have to request that any submission will have references to reputable sources.

    • Sentient_Modem@lemm.ee
      link
      fedilink
      arrow-up
      2
      ·
      3 个月前

      I would probably start out by proofing or approving them before they post to the site. It say I get a notification read it do a little reading over it and get to a point where I can use a large language model to siphon the submissions.