• KeenFlame@feddit.nu
    link
    fedilink
    English
    arrow-up
    2
    ·
    13 hours ago

    Maybe I misunderstood, are you saying all hallucinations originate from the safety regression period? Because hallucinations appear in all architectures of current research, open models, even with clean curated data included. Fact checking itself works somewhat, but the confidence levels are off sometimes and if you crack that problem, please elaborate because it would make you rich

    • 𞋴𝛂𝛋𝛆@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 hours ago

      I’ve explored a lot of patterns and details about how models abstract. I don’t think I have ever seen a model hallucinate much of anything. It all had a reason and context. General instructions with broad scope simply lose contextual relevance and usefulness in many spaces. The model must be able to modify and tailor itself to all circumstances dynamically.