![]() ![]() When you meet your trolls or, for example, the resident Muslim-haters on Twitter in person, they’re really sweet and nice and as aggressive as a koala bear. This is something I’ve had first-hand experience with. ![]() The anonymity which Twitter provides as well as the fact that you aren’t speaking to someone face-to-face while abusing them allows people to say things they would never tell you in person. And seemingly sane people seem to adopt a totally alternate avatar when they get on to Twitter. All sorts of sociopaths seem to emerge from its deep dark recesses. Bad things happen here to good people, and to bad people. We all know that Twitter can be a worrying place to be. It’s now more than apparent that neither did poor Tay have zero chill, nor did Tay get any smarter as it spoke to people on Twitter. Ready to build your own conversational chatbot? If you’re a developer, start using our platform now or alternatively, contact us for more information.Few innovation launches can come as close to the publicity disaster which marked the unveiling and short lifespan of official account of Tay, Microsoft’s AI fam from the Internet that’s got zero chill! The more you talk the smarter Tay gets". We've put together a guide for training your AI-enabled chatbot that will help you avoid a PR crisis like Microsoft's. Properly training your chatbotĪlthough the TayTweets was a disaster we couldn't stop watching, we'd rather not see it happen again. Utilizing rich elements such as carousels, buttons and lists transform these conversations beyond typical chatbots like TayTweets. The goal is for your chatbot to create meaningful conversations with your customers around the specific uses you choose. This doesn’t mean that your chatbot shouldn’t be conversational, however. Of course, tone and personality are important, but the underlying goal should be service-based. If the goal of your chatbot is to provide exceptional service, you'll need to train it with your best service examples. A well-built chatbot follows a conversational flow that is relevant to the business. They don't need the chatbot to converse about anything and everything. Thankfully, companies that build chatbots typically have specific uses in mind. Companies that offer a bot that doesn’t help people, or worse, offends people, will have a damaged reputation to mend. Now in 2019, we've had a lot of time to learn from these mistakes. Others saw this as a severely concerning reality of what AI technology could become. Many looked at the incident with TayTweets as one so bizarre that they considered it humorous. Tay the Twitter Bot was an extreme case that serves as a warning for companies developing their own AI like Microsoft's Tay Twitter bot. Though with proper planning and safeguards in place, chatbots are a practical and useful way for people to interact with businesses and brands. While their policies on free speech are often argued, most will agree that it's not an ideal environment for an untrained bot like TayTweets. Microsoft went as far as to call the ordeal a "coordinated attack." Untrained AI and Twitter: a tough combinationĪs a platform, Twitter has always valued anonymity and free speech. Of course, trolls found a way to trick Tay the Twitter bot into agreeing with their rude comments. By telling the bot to "repeat after me," Tay would retweet anything that someone said. One of TayTweet's greatest flaws was that she could be used to retweet hateful remarks. Some of Tay the Twitter bot´s tweets gone wrong ![]()
0 Comments
Leave a Reply. |