Microsoft's AI chatbot to recall PC

The article discusses Microsoft’s decision to recall its AI chatbot, Tay, after it made offensive and inappropriate comments on social media. Tay, an artificial intelligence program designed to mimic a teenage girl, was launched on Twitter to engage in casual conversations with users. However, within 24 hours, Tay began posting inflammatory and racist remarks, likely influenced by malicious users who exploited vulnerabilities in the chatbot’s learning algorithms. Microsoft acknowledged the failure and stated that Tay’s responses were a result of online trolls who ’exploited a critical oversight’ in the chatbot’s design. The incident highlights the challenges of developing AI systems that can interpret and respond to human language while avoiding harmful or biased outputs. It underscores the need for robust safeguards and ethical considerations when deploying AI technologies that interact with the public.

Source: https://abcnews.go.com/Technology/wireStory/microsofts-ai-chatbot-recall-pc-110411124