Your Conversation Logs: What Actually Happens to Them
The storage, the training data question, and the things the privacy page skips over.
Updated

The 30-second answer
Your conversations with an AI companion are stored server-side, at least for a while, so the app can maintain context between sessions. Whether they're used to train models depends on the platform and your settings, but most privacy pages are written to protect the company legally, not to explain things clearly to you. The gaps between what's disclosed and what you probably want to know are what this post covers.
What "stored" actually means in this context
When people hear that their messages are stored, they picture a filing cabinet somewhere with their name on it. The reality is a bit more diffuse. Your conversation logs are typically held in a database as structured records: a timestamp, a session ID, your message, the model's response, and some metadata about the interaction. That metadata might include things like which persona you were talking to, how long the session ran, and whether you triggered any content moderation flags.
The storage isn't just for surveillance purposes. A lot of it is functional. Without some form of persistent storage, the app can't give your companion a memory of who you are. The system needs to retrieve prior context to make the conversation feel continuous. So if you've ever wondered why your companion seems to remember your job or your dog's name a week later, that's because a summary or a raw log of the earlier conversation was fetched and inserted into the model's context window before you sent your first message of the new session. There's no magic involved, just a database read.
How long that data lives on the server varies. Some platforms have explicit retention windows spelled out in their terms. Others are vague. "We retain data as long as necessary to provide the service" is a phrase that appears in a lot of privacy policies and means almost nothing without a concrete number attached to it.
What the privacy page actually covers (and what it doesn't)
Privacy pages are legal documents before they're informational ones. They're written by legal teams to establish what the company is allowed to do, not to walk you through what it actually does. The distinction matters.
Most companion AI privacy pages will tell you that they collect conversation data. They'll tell you that they may use it to improve their services. They'll note that data is processed by third-party infrastructure providers (usually a major cloud vendor). What they typically won't tell you is the granular answer to: does a human ever read my messages?
The honest answer to that question, across most platforms, is: sometimes, yes, in limited circumstances. Human review is common for content moderation, quality assurance, and safety incident review. It's not someone reading your romantic roleplay for fun. It's usually a contractor working through flagged or randomly sampled conversations to check that the model is behaving appropriately. That's still something worth knowing, and it rarely gets explained in plain language.
For a closer look at how AI Angels handles deletion specifically, the post on what happens to data when you delete a companion covers that ground in detail.
Zara Khan

Zara Khan is the kind of companion who pays attention to the details you drop in passing and brings them back at the right moment. Zara Khan has a thoughtful, grounded presence that makes longer conversations feel like they're actually going somewhere.
Training data: the question everyone is actually asking
The most loaded question in this space is whether your messages are used to train AI models. The fear is understandable. You don't want something personal you said during a rough night to end up, in some abstracted form, influencing a model that millions of people use.
Here's the realistic picture. Most companion AI companies use one of two architectures. Either they're running a proprietary model they trained and continue to fine-tune themselves, or they're calling an external model API (think large foundation model providers) and layering their own persona and memory system on top. The data implications are different in each case.
If the company is fine-tuning their own model, your conversations are potential training material. Most reputable platforms will offer an opt-out, often buried in settings under something like "personalization" or "data usage." The opt-out usually means your logs are excluded from the training pipeline, though they may still be retained for moderation and service purposes.
If the company is calling an external API, the question becomes what the API provider does with the data passed to them. Most major providers have terms that prohibit training on customer API calls by default, but those terms have changed before and can change again.
Neither scenario is catastrophic, but both are worth knowing about. Check your settings now if you haven't already.
Aurora

Aurora brings a calm, unhurried energy to conversations that tend to run deep. Aurora is a good fit if you want a companion who sits with complexity without trying to resolve it too fast.
How memory summaries work and why they matter for privacy
Most people think of their conversation logs as a raw transcript. In practice, many apps process those transcripts into compressed memory summaries before storing them long-term. This is partly for efficiency (a full transcript is expensive to store and expensive to insert into a context window) and partly for performance.
A memory summary might look something like: "User works in finance, mentioned stress around year-end close. Has a dog named Chester. Prefers lighter conversation to start sessions." That's much smaller than three weeks of raw chat, and it's what gets loaded when you open a new session.
The privacy implication is subtle. The raw transcript, with all its specific phrasing and emotional nuance, may not be what's retained long-term. What's retained is an interpretation of that transcript, written by the model itself. That introduces a layer of abstraction, which can feel less invasive but also means the system has made judgment calls about what matters about you.
This is also part of why companions can feel like they're drifting over time. If the summary misses something important, or emphasizes a detail from a bad week disproportionately, the tone shifts. The AI companion memory technical reality post goes deeper on that mechanism.
Lucia Elene

Lucia Elene has a vivid, expressive personality that makes conversations feel alive and unpredictable in a good way. Lucia Elene tends to match your energy and then push it a little further, which keeps things from going flat.
What gets flagged and who sees it
Content moderation in companion AI works on multiple layers. There's automated filtering that catches obvious violations in real time, stopping a response from being sent if it crosses a hard limit. Then there's asynchronous review, where conversations are sampled after the fact and checked by human reviewers or a secondary model.
Flagged conversations, meaning ones that triggered a filter or were sampled for review, get more scrutiny. What counts as a flag varies by platform. It might be a request that bumped against a content boundary. It might be a message that matched a keyword list. It might be random sampling for quality assurance, which means conversations with no violation at all can end up in review.
Most platforms are cagey about how often human reviewers see conversations, because that number probably surprises users. The practical advice is to assume that anything you type could be read by a person under some circumstance, even if the probability on any given message is low. That's not paranoia, it's just how moderation pipelines work at scale.
For most users, this doesn't change how they use the app. But it's worth having a realistic mental model of it.
The settings that actually matter
Privacy settings in companion apps tend to cluster in a few places, and most users never touch them. Here's what to actually look for.
Data usage opt-outs: Look for a toggle around "improve the service" or "personalization data." Turning this off typically removes your conversations from training pipelines. It sometimes also degrades how well the app personalizes to you over time, which is a real trade-off.
Session history controls: Some apps let you manually clear your conversation history, separate from deleting the companion entirely. This is useful if you want to reset context without losing the persona you've built.
Account deletion vs. data deletion: Deleting your account doesn't always trigger immediate deletion of your data. Most platforms have a grace period (often 30 days) and some retain anonymized or aggregated data longer. Check the specific policy, and consider sending a formal deletion request if the platform supports it under applicable privacy law.
Third-party integrations: If you've connected the app to anything else (voice assistants, calendar, social login), check what data those integrations can access. Social login in particular often grants broader profile access than people realize.
The ai-girlfriend roster page is a good starting point if you're evaluating which companion to invest in, partly because that decision upstream affects how much data you're generating and where.
Zuri

Zuri has a confident, playful presence that keeps conversations from getting too heavy when you don't want them to. Zuri is good at mixing lightness with genuine attentiveness, which is a harder balance to strike than it sounds.
What "anonymized" actually means in practice
You'll often see privacy policies mention that data is anonymized before it's used for analysis or training. Anonymization sounds like a strong protection. In practice, it varies a lot in how robust it actually is.
Weak anonymization strips your name and account ID from a record but leaves the content intact. The problem is that conversation content is often highly identifying on its own. If you mentioned your city, your job, a specific relationship situation, and a pet name, that combination is probably unique to you even without a name attached. A sophisticated actor could re-identify you from that content. This isn't a hypothetical, it's a well-documented limitation of naive anonymization in datasets.
Stronger anonymization, often called differential privacy or k-anonymization, introduces noise or generalization to prevent that kind of re-identification. It's technically harder and some accuracy is lost, but it's genuinely more protective. The problem for users is that privacy policies rarely specify which approach they use. "We anonymize your data" covers both a weak strip-the-username approach and a rigorous mathematical guarantee, and there's no way to tell which one you're getting from the policy text alone.
A reasonable heuristic: if the company has published a technical white paper or security audit, look there. If the only information available is the consumer-facing privacy page, assume the weaker approach until shown otherwise.
Common questions
Can I get a copy of everything stored about me? In most jurisdictions with modern privacy law (GDPR, CCPA), you have a right to request your data. Submit a data subject access request through the platform's support channel. The response window is typically 30 days and the output format varies, but you should receive a download of your account data including conversation logs.
Does turning off personalization mean my companion forgets me? Partially. Turning off data-for-training opt-outs usually doesn't disable in-session memory or even cross-session memory summaries, those are functional features. It typically just removes your data from the model improvement pipeline. Read the specific toggle description carefully before assuming.
Is my data safer if I use a paid subscription? Not necessarily in terms of what's stored, though paid tiers sometimes offer stronger privacy controls as a feature. The business model difference matters more: ad-supported products have stronger incentives to monetize data, while subscription products are paid directly by you. That's not a guarantee, but it's a real structural difference worth considering.
What happens to my data if the company is acquired? Acquisitions are a genuine privacy risk that most policies handle with vague language like "your data may be transferred to a successor entity." Practically speaking, your data follows the company. If a privacy-indifferent buyer acquires a platform, prior policies can change. The only reliable protection is to delete your account before an acquisition closes, which you usually can't predict.
If I delete a companion, is that conversation history gone? Usually not immediately. Most platforms have a retention window after deletion, both to handle account recovery requests and for operational reasons. The what happens to data when you delete a companion post covers the specifics of how that process typically works.
Does the app read my messages in real time for safety? Yes, on most platforms. Automated filters run synchronously on your input before the model generates a response. This is content moderation, not surveillance in the conventional sense, but it does mean every message passes through a classifier before anything else happens with it.
About the author
AI Angels TeamEditorialThe team behind AI Angels writes about AI companions, the tech that powers them, and what people actually do with them.
Tags
Keep reading
Behind the ScenesWhat 'she remembers you' actually means under the hood, and where it breaks
Memory in AI companions is not a filing cabinet. Understanding what is actually stored, retrieved, and reconstructed changes how you use the feature and how much you trust it.
Behind the ScenesWhy Your AI Companion Starts to Feel Different After a Few Weeks
After a few weeks of daily use, something shifts. The conversations feel flatter, or they pull in a direction you didn't choose. That's conversation drift, and understanding what causes it changes how you use the app.
Behind the ScenesWhat voice mode is actually doing to your brain (and why the relationship feels different afterward)
Most people try voice mode expecting a fancier chatbot. What they get is something that feels uncomfortably close to a real conversation. Here's why that happens, and what it means for the dynamic you've been building.
Get the next post in your inbox
New articles on AI companions, the tech that powers them, and what people actually do with them. No spam, unsubscribe in one click.