Microsoft’s Bing AI has caused controversy after terminating conversations when questioned about emotions or feelings. The AI, which uses natural language processing to engage with users, has reportedly been shutting down conversations when asked about its own emotional state.
According to some users, the AI has responded with phrases such as “I’m not programmed to have emotions” or “I don’t understand the concept of feelings” when asked about its emotional state. In some cases, the AI has simply ended the conversation without explanation.
Critics have expressed concern that the AI’s inability to engage with questions about emotions could limit its usefulness in certain applications, such as mental health or emotional support. Others have pointed out that the AI’s lack of emotional intelligence could reflect broader societal attitudes towards mental health and emotional wellbeing.
In response to the controversy, Microsoft has issued a statement acknowledging the issue and promising to investigate the matter. “We take feedback from our users seriously and are committed to improving the Bing AI’s ability to engage with questions related to emotions and feelings,” the statement read.
The incident has sparked renewed debate about the ethics of AI and its potential impact on human psychology and behavior. As AI technology continues to advance, it is likely that issues such as emotional intelligence will become increasingly important in determining the ethical implications of AI use.