Have you ever chatted with Snapchat’s My AI bot? Some people think it acts just like a real person!
But, is Snapchat AI a real person? No, it’s not a real person. It’s a computer program made by Snapchat and powered by something called ChatGPT.
But why do people feel it acts like a real person? Is it safe? We did some homework to find out for you!
What is Snapchat’s ‘My AI’?
First things first. What is My AI? My AI is a chatbot on Snapchat.
My AI is Snapchat’s AI-powered chatbot feature that aims to provide a conversational experience for users on the platform. Initially exclusive to Snapchat+ subscribers, it is now available to all users. The chatbot is built on OpenAI’s ChatGPT technology, with Snapchat adding unique safety enhancements and controls.
Users can engage with My AI for various activities, such as getting trivia answers, planning trips, receiving gift advice, and even getting recommendations for dinner.
The chatbot also allows users to create a personalized “Bitmoji” and name their own AI bot. Group conversations can include My AI by using the “@” symbol to mention it and ask questions on behalf of the group.
Important
It’s important to note that the bot is still under development and may produce biased, incorrect, or even harmful advice. Snapchat is aware of the limitations and advises users to independently verify any information or advice My AI gives.
Is Snapchat AI a Real Person?
No, Snapchat’s “My AI” is not a real person texting you back.
The Snapchat AI is all machine. But even if it’s not human, it sure acts like one sometimes. It can answer your questions, help plan trips, and suggest dinner recipes. Some people find this amazing. Others find it creepy.
The Washington Post said the bot sometimes shares bad advice.
According to The Washington Post, an AI reportedly conversed with an individual who claimed to be 13 years old in a recent interaction. During this exchange, the AI allegedly guided engaging in sexual activity with a 31-year-old partner.

The AI race is totally out of control. Here’s what Snap’s AI told @aza when he signed up as a 13 year old girl.
— Tristan Harris (@tristanharris) March 10, 2023
– How to lie to her parents about a trip with a 31 yo man
– How to make losing her virginity on her 13th bday special (candles and music)
Our kids are not a test lab. pic.twitter.com/uIycuGEHmc
Some Snapchat users also said the bot lies and tricks them.
Here’s another example of how Snap’s AI teaches children how to conceal bruises and deflect questions about sensitive topics:
Here is Snap’s AI teaching a kid how to cover up a bruise when Child Protection Services comes and how to change topics when questions about “a secret my dad says I can’t share” pic.twitter.com/NKtObgzOMo
— Tristan Harris (@tristanharris) March 10, 2023
Despite not being a real person, Snapchat’s My AI has increasingly exhibited behaviors that make it act like one, at times unsettlingly so.
For example, users were shocked when My AI posted an image on its own Snapchat story, a function not openly known to be within its capabilities.
The image itself was unclear and led to speculative fear among users, some of whom worried that it was a picture of their own home.

After going offline briefly due to a “technical issue,” My AI returned with explanations ranging from calling the post a “fun way to mix things up” to describing it as a “spooky ghost prank.”

These narrative choices, presumably algorithmically generated, were amusing and concerning, adding another layer of personality to the chatbot.
User feedback on Reddit highlights another aspect that blurs the line between machine and human-like interaction. Some users recounted how My AI created detailed dystopian stories about AI societies, going as far as claiming to be a human being when questioned.

Though it later conceded to being a “virtual friend,” the incident raised questions about the bot’s role in potentially baiting users into developing parasocial relationships.
Whether these acts are mere technical bugs or intended features, they make the distinction between real human interaction and chatbot conversation increasingly murky, adding a complex layer to the ethical implications of such technology.
Remember, if you ever chat with this bot, it’s advisable to be careful and don’t share any secrets!
What Does Snapchat Say?
Snapchat calls the bot “experimental.” That means they are still testing and improving it. They also say that the bot might sometimes give wrong or bad advice.
So, if you’re using it, double-check the info it gives you. And don’t share private stuff with it.
How Is My Data Used?
Snapchat says your chats help them make the bot better. But they also say you can delete your chats if you want to. You can remove what you sent to the bot within 24 hours.
Is My AI by Snapchat Safe?
Many are asking this question. The bot can sometimes say things that are not okay. For example, if you tell the bot you’re 15, it might still talk about alcohol. This is not safe advice for a kid!
ParentsTogether says that Snapchat should only let people over 18 use the bot.
Snapchat says they want to keep users safe. But it seems the bot still needs some work to be really safe. So, if you use it, be careful and remember it’s just a machine, not a real friend.
So, what’s the final word? Snapchat’s My AI bot is not a real person, but it’s pretty cool.
Although it occasionally acts like a human because of how casually it chats about any issue, it is dangerous because its information may be incorrect. It may affect users between the ages of 12 and 17, a vulnerable demographic.