How to Use AI Without Giving Away Your Data
ChatGPT and Gemini collect your conversations. Here's how to use AI privately — self-hosted, API-only, EU data residency, and why the lobsterfarm model keeps your data yours.
How to Use AI Without Giving Away Your Data
TL;DR: When you use ChatGPT or Gemini through their web apps, your conversations become their training data (unless you opt out, and even then it's murky). When you use AI through an API with your own key, providers explicitly say they don't train on your data. Self-host on your own server and the data never leaves your machine. Here's the full picture.
The Problem
Every time you type something into ChatGPT's web interface, you're doing two things:
- Getting a helpful AI response
- Giving OpenAI your data
OpenAI's terms of service for consumer products (ChatGPT free and Plus) state that conversations may be used to improve their models. Google's Gemini terms are similar. This means:
- Your business ideas? Training data.
- Your personal problems? Training data.
- That contract you asked it to review? Training data.
- Your code with proprietary business logic? Training data.
Yes, you can turn off "model training" in ChatGPT's settings. But there are caveats:
- It's not the default — you have to know to look for it
- It's unclear exactly what "not used for training" means in practice
- Your data still lives on their servers, subject to their retention policies
- Human reviewers may still see your conversations for safety purposes
- The setting has changed multiple times — what you agreed to last year might not be what's in effect today
This isn't a conspiracy theory. It's the business model. Free and cheap AI products are subsidized by the value of the data they collect.
The Three Levels of AI Privacy
Level 1: Consumer Products (Least Private)
What: ChatGPT web/app, Gemini web/app, Copilot free tier
Your data: Stored on provider servers, potentially used for training, subject to employee review, retained according to their policies (which change).
Good for: Casual use, non-sensitive topics, things you'd be fine posting publicly.
Not good for: Business strategy, legal documents, personal health info, financial details, proprietary code, anything you wouldn't want a stranger to read.
Level 2: API Access (Much Better)
What: Using AI through API keys with tools like OpenClaw
Your data: Both Anthropic and OpenAI have explicit, legally binding policies for API usage:
Anthropic (API Terms): "We do not train our models on Customer Content submitted via the API."
OpenAI (API Terms): "We do not use data submitted through the API to train or improve our models, unless you explicitly opt in."
This is a fundamentally different deal than the consumer products. API data policies exist because enterprise customers (banks, hospitals, law firms) demand it. The legal and financial consequences of violating these terms would be severe.
Your data is still sent to the provider's servers for processing. They see it temporarily. But they don't keep it for training, and retention policies are shorter and more clearly defined.
Good for: Most use cases. Business, personal, professional. The privacy guarantee is strong enough for serious work.
Not good for: Situations where data absolutely cannot leave your network (classified info, extreme regulatory environments).
Level 3: Self-Hosted / On-Premises (Most Private)
What: Running a local AI model on your own hardware. No data leaves your machine.
Your data: Never touches the internet. Never seen by any company. Exists only on hardware you physically control.
Good for: Maximum privacy requirements, air-gapped environments, classified or extremely sensitive data.
Trade-offs: Local models are less capable than cloud models (Claude Opus, GPT-4). You need decent hardware ($500-2,000 for a capable GPU). Setup and maintenance is on you.
OpenClaw supports local models for exactly this use case — full privacy with the same assistant experience, just a less powerful brain.
The API-Only Path: Best Balance for Most People
For the majority of users, API access is the sweet spot. You get world-class models (Claude, GPT-4) with a clear, legally binding guarantee that your data isn't used for training.
Here's what the API path looks like in practice:
- Get an API key from Anthropic or OpenAI ($0 to set up, pay only for usage)
- Run OpenClaw on a server or through lobsterfarm
- Chat through your messaging app (Telegram, WhatsApp, Discord)
Your message flow:
Your phone → Your server (OpenClaw) → AI Provider API → Your server → Your phone
At no point does your data sit in a consumer product's training pipeline. The provider processes your request and returns a response. That's it.
EU Data Residency: Why Location Matters
Where your data is processed matters, especially under GDPR and similar regulations.
Anthropic: US-based company with US data centers. When you use Claude's API, your data is processed in the US. There's a pending EU region but it's not available yet.
OpenAI: US-based company with US data centers. Same situation.
Your server: This is what you control. If you run OpenClaw on a server in Germany or Finland, your memory files, conversation history, and personal data stay in the EU. Only the individual API calls cross the Atlantic — and those are transient, not stored.
The Hetzner Advantage
Hetzner runs data centers in Germany (Falkenstein, Nuremberg) and Finland (Helsinki). They're popular with privacy-conscious users because:
- Servers physically located in the EU
- German company, subject to German data protection law
- Excellent pricing (€4-8/month for a capable VPS)
- No data mining, no advertising, no hidden data collection
When you run OpenClaw on a Hetzner server:
- Your memory files (MEMORY.md, SOUL.md, daily notes) → stored in Germany/Finland
- Your conversation history → stored in Germany/Finland
- Your personal data → stored in Germany/Finland
- API calls → temporarily processed by Anthropic/OpenAI in the US, but not retained
This gives you the best of both worlds: world-class AI models with EU-resident persistent data.
See our Hetzner setup guide for step-by-step instructions.
The lobsterfarm Model
lobsterfarm's managed OpenClaw instances run on Hetzner infrastructure in Germany and Finland. Here's the privacy architecture:
Your server: Isolated instance, just for you. No shared resources with other users. Your data lives on your dedicated instance in the EU.
Your API key: You bring your own Anthropic or OpenAI key. We never see your API traffic. The connection goes directly from your instance to the provider.
Your data: We don't have access to your conversations, memory files, or personal data. We can access server logs for support purposes, but conversation content is yours.
Your control: You can export everything at any time — configs, memory files, conversation data. You can delete your instance and all data is gone.
Think of it this way: lobsterfarm is like a landlord. We provide the apartment (server) and keep the building running (infrastructure), but we don't read your mail (conversations) or go through your closet (data).
Practical Privacy Tips (For Any AI Setup)
Whether you use lobsterfarm, self-host, or even stick with ChatGPT, these habits improve your AI privacy:
1. Use API Access, Not Consumer Products
If you're going to have deep, ongoing conversations with an AI, use the API path. Consumer products are fine for quick one-off questions. Ongoing personal or business use deserves the API's privacy guarantees.
2. Be Intentional About What You Share
Even with strong privacy guarantees, think about what you're sending to a third-party API. You probably don't need to send your social security number or credit card number through any AI. Common sense applies.
3. Set Up Data Retention Policies
If you're self-hosting, decide how long to keep conversation logs. Old logs are a liability — if your server is compromised, less data means less exposure. Set up automatic log rotation.
4. Encrypt Your Disk
If you're running a VPS, enable full-disk encryption. If the physical hardware is compromised, your data remains protected.
5. Use Separate API Keys
Don't use the same API key for personal use, work projects, and shared tools. Separate keys mean separate billing, separate usage tracking, and easier auditing.
6. Review Your Memory Files
OpenClaw's memory system stores things you've told it. Periodically review MEMORY.md and daily notes. Remove anything you don't want persisted. These files are plain text — you have complete visibility and control.
7. Understand the Limits
No system is perfectly private. API calls transit the internet. Providers retain logs for some period. Server breaches happen. The goal isn't perfection — it's raising the bar significantly above "typing your secrets into a consumer product that mines your data."
The Privacy Spectrum
Here's a realistic ranking of AI privacy options:
| Setup | Privacy Level | Convenience | Cost |
|---|---|---|---|
| Local model, air-gapped | 🟢🟢🟢🟢🟢 | 🔴 | Hardware costs |
| Self-hosted + API key | 🟢🟢🟢🟢 | 🟡 | VPS + API |
| lobsterfarm (managed) | 🟢🟢🟢🟢 | 🟢🟢 | Service fee + API |
| ChatGPT API (direct) | 🟢🟢🟢 | 🟡 | API only |
| ChatGPT Plus (training off) | 🟢🟢 | 🟢🟢🟢 | $20/month |
| ChatGPT Free | 🟢 | 🟢🟢🟢🟢 | Free |
Most people should be in the middle of this spectrum: API access with a server they control (or a managed service like lobsterfarm). You get 90% of the privacy benefit with 90% of the convenience.
The Bottom Line
You don't have to give away your data to use AI. The API-only path gives you world-class models with strong, legally binding privacy guarantees. Self-hosting adds another layer. EU hosting adds data residency compliance.
The key insight: privacy is a configuration choice, not a feature you pay extra for. The same AI models are available through consumer products (low privacy) and APIs (high privacy). You just need to choose the right path.
lobsterfarm runs on EU infrastructure with BYO API keys, isolated instances, and full data export. Your conversations stay on your server, your data stays in the EU, and you stay in control.
Get started with lobsterfarm → · Self-host on Hetzner → · Learn about API costs →
Skip the setup. Start using your AI assistant today.
lobsterfarm gives you a fully managed OpenClaw instance — one click, your own server, running 24/7.