• kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    6 months ago

    Basically, any time a user prompt homes in on a concept that isn’t represented well in the AI model’s training dataset, the image-synthesis model will confabulate its best interpretation of what the user is asking for.

    I’m so happy that the correct terminology is finally starting to take off in replacing ‘hallucinate.’