• SavvyWolf@pawb.social
    link
    fedilink
    English
    arrow-up
    4
    ·
    4 days ago

    “Hate” and “Love” are complex things that are very ingrained into human nature. For an AI or robot to have these things, it would essentially have to emulate or implement a human mind. Such a thing is currently very far beyond our current technology level, and arguably isn’t even a goal of many AI projects. Most AI systems like ChatGPT are basically glorified autocomplete. They are given an input and use complex maths and probabilities to predict what a human would respond to it. They don’t have any understanding about what they are talking about.

    I think if an AI were able to hate or love, it would raise complicated and perhaps uncomfortable questions about what it means to be “human”. Can a system that perfectly replicates human emotions and experiences, not be considered human itself?