The AI question every product team is asking
If you are building or maintaining an iOS app in 2026, someone on your team has already asked: "Should we add AI?" Maybe it was your CEO after a board meeting, a product manager after a competitor launch, or a customer support lead who is tired of answering the same questions.
The question is valid. The problem is that most teams jump straight to implementation without asking the harder questions first: What kind of AI? How much will it cost per user per month? Does it actually solve a problem our users have? And can we do it without violating GDPR?
This post is the decision framework I wish someone had given me before I integrated AI into Second Brain, our iOS memory companion. It is written for founders, CTOs, and product leads — not developers. No code, no technical jargon. Just the decisions you need to make, what each option costs, and where the hidden traps are.
Two paths: on-device AI vs. cloud AI
The first fork in the road is where the AI runs. This single decision shapes your cost structure, privacy posture, and user experience more than anything else.
On-device AI
Apple now provides Foundation Models that run directly on the user's iPhone or iPad — no server, no API call, no data leaving the device. These models handle tasks like text classification, summarization, entity extraction, and basic generation. They are free to use (zero API cost per request), work offline, and respond in milliseconds.
The trade-offs are real, though. On-device models are smaller and less capable than cloud models. They cannot handle complex reasoning, long-form generation, or nuanced multi-step analysis. They require recent hardware (typically iPhone 15 Pro or newer for the larger models), which limits your addressable market. And because Apple controls the model, you cannot fine-tune it or swap it for something better.
Best for: classification, tagging, simple text analysis, auto-complete, spam filtering, on-device search ranking. Tasks where speed and privacy matter more than sophistication.
Typical cost: Zero per-request cost. Development effort is 2 to 4 weeks for a senior iOS developer to integrate and test.
Cloud AI (API-based)
Cloud models like Anthropic's Claude, OpenAI's GPT, or Google's Gemini run on remote servers. Your app sends a request, the model processes it, and returns a response. These models are far more capable — they can write, reason, analyze documents, and handle complex conversational interactions.
The cost model is pay-per-use: you pay for every request based on the number of tokens (roughly, words) sent and received. This means your AI cost scales directly with user engagement. A popular feature that users love can also be the feature that blows your monthly budget.
Best for: conversational interfaces, complex analysis, content generation, personalized recommendations, anything requiring reasoning or large context windows.
Typical cost: $0.001 to $0.05 per request depending on model tier and prompt length. Development effort is 4 to 8 weeks including prompt engineering, error handling, and cost controls.
The hybrid approach (and why it often wins)
In Second Brain, we use both. Apple's on-device Foundation Models handle the fast, cheap tasks: classifying new memories into categories, extracting key entities, tagging content. This happens instantly with zero API cost. For the features that need deeper intelligence — exploring connections between memories, generating weekly reflections — we call Claude's API, routing between Haiku (faster, cheaper) for exploratory queries and Sonnet (more capable) for reflections that need nuance.
This hybrid approach keeps our average API cost per active user under $0.30 per month, while still offering genuinely intelligent features. If we had used a cloud model for everything, that number would be closer to $2.00 per user per month — unsustainable for a consumer app with a $4.99 monthly subscription.
Decision framework: Use on-device AI for anything that can be done locally with acceptable quality. Reserve cloud AI for the features that genuinely need it. The boundary is not technical — it is economic. Every cloud request is a line item on your P&L.
What it actually costs: a realistic breakdown
The total cost of adding AI to your iOS app falls into three buckets: development, ongoing API costs, and infrastructure. Most teams underestimate the second and ignore the third.
Development costs
For a basic AI feature (single model, one use case, no conversation history), expect 4 to 6 weeks of an experienced iOS developer's time. That includes API integration, prompt engineering, error handling, loading states, offline fallbacks, and testing across devices and network conditions.
For a more sophisticated integration (multiple models, conversation context, hybrid on-device/cloud, cost tracking per user), plan for 8 to 12 weeks. This is what Second Brain required. The API integration itself was perhaps 20% of the effort. The other 80% was prompt iteration, edge case handling, cost controls, and making the feature feel fast on slow connections.
At typical European freelance rates for senior iOS developers (EUR 80 to 150 per hour), you are looking at:
- Basic AI feature: EUR 12,000 to 36,000
- Sophisticated AI integration: EUR 25,000 to 72,000
These ranges are wide because the actual cost depends heavily on your app's existing architecture, how well-defined the AI use case is, and whether you need conversation history, user-specific context, or multi-model routing.
Ongoing API costs
This is where most business plans get it wrong. Teams model API costs based on average usage, but usage patterns for AI features are not average — they follow a power-law distribution. A small percentage of power users will generate 10 to 50 times more API calls than the median user.
Realistic ranges for a consumer iOS app (per active user per month):
- Light AI features (auto-tagging, smart search): $0.01 to $0.10
- Moderate AI features (chat-like interface, analysis): $0.20 to $1.00
- Heavy AI features (document processing, long conversations): $1.00 to $5.00+
For a B2B app with fewer but more intensive users, multiply these by 5 to 10. The numbers get meaningful fast: 10,000 active users at $0.50 per month is $5,000 per month in API costs alone — before you have paid for anything else.
Cost trap: Many teams forget to implement per-user rate limits and spending caps before launch. Without them, a single power user or a bot can generate hundreds of euros in API charges in a single day. Build cost controls into your architecture from day one, not as a post-launch patch.
Infrastructure costs
If you call an AI API directly from the user's device, your infrastructure cost is nearly zero — you just need an API key management strategy. But most production apps need a thin backend layer between the device and the AI provider, for rate limiting, usage tracking, API key rotation, and response caching. Budget EUR 50 to 200 per month for a lightweight server or serverless functions to handle this.
GDPR and privacy: the European dimension
If your users are in Europe — and if you are reading this from Helsinki, Dublin, or Berlin, they almost certainly are — GDPR creates specific constraints on how you can use AI.
On-device AI is your safest option
Data that never leaves the device is data you do not need to worry about in your privacy impact assessment. On-device processing using Apple's Foundation Models means no data transfer, no third-party processor agreements, and no cross-border data concerns. From a GDPR perspective, it is almost invisible.
Cloud AI requires careful handling
When you send user data to a cloud AI provider, you are transferring personal data to a third-party processor. This means you need:
- A Data Processing Agreement (DPA) with your AI provider. Both Anthropic and OpenAI offer these — review them carefully, especially the data retention clauses.
- A lawful basis for processing. For most apps, this will be either consent or legitimate interest. If your AI feature is core to the app's function, legitimate interest may apply. If it is optional or experimental, explicit consent is safer.
- Transparency in your privacy policy. Users must know their data is being sent to an AI service, which service, and what it is used for. Vague language like "we may use AI to improve your experience" is not sufficient.
- Data minimization. Send only what the AI model needs. If you are asking Claude to classify a memory entry, send the text — not the user's name, location, or full history.
In Second Brain, we took a belt-and-suspenders approach. The privacy policy clearly states that text is sent to Anthropic's Claude API for specific features. We use Anthropic's zero-retention API option so that user content is not stored on their servers after processing. And we minimize what we send — the AI sees individual memory entries, never the user's full history in a single request.
The US data transfer question
Most major AI providers are US-based. Under the EU-US Data Privacy Framework (which replaced Privacy Shield), transfers to certified US companies are permitted. But the legal landscape shifts frequently. If regulatory stability matters to your risk assessment, on-device processing or a European AI provider eliminates this concern entirely.
When AI adds value vs. when it is a gimmick
Not every app needs AI. Adding it because competitors have it, or because investors expect it, is a fast path to wasted budget and a cluttered product.
AI adds genuine value when:
- It saves users meaningful time. Auto-classifying 50 daily entries that a user would otherwise tag manually. Generating a weekly summary from scattered notes. These solve real friction.
- It reveals patterns humans miss. Finding connections between symptoms logged over months. Identifying spending patterns across hundreds of transactions. The value comes from scale, not intelligence.
- It makes a complex interface accessible. Letting a user ask "What did I capture about project Alpha last month?" instead of scrolling through a timeline. Natural language as navigation.
AI is likely a gimmick when:
- A simple rule would work just as well. If your "AI-powered" feature is really just an if-else statement with a language model wrapper, save the API costs.
- Users did not ask for it. Adding a chatbot to a utility app because the technology exists is not a product strategy. Ask your users what slows them down, then evaluate whether AI is the right tool.
- The accuracy is not high enough to trust. An AI feature that is right 80% of the time might sound impressive in a demo but creates frustration in daily use. If users have to verify every AI suggestion, you have added work instead of removing it.
Litmus test: Before committing budget, describe the AI feature without using the word "AI." If the description still sounds valuable — "automatically categorizes entries and surfaces weekly trends" — it is probably worth building. If it only sounds impressive because of the technology — "uses a large language model to..." — reconsider.
A real-world case study: Second Brain
Second Brain is an iOS app that helps users capture and recall thoughts using AI. Here is how the economics work in practice:
- On-device AI (Apple Foundation Models): handles memory classification, entity extraction, and quick tagging. Zero API cost. Covers roughly 70% of all AI interactions in the app.
- Cloud AI (Claude API via Anthropic): handles explore queries (Haiku model, fast and cheap) and weekly reflections (Sonnet model, more capable). Covers the remaining 30% of AI interactions.
- Average API cost per active user: under $0.30/month.
- Subscription price: $4.99/month (Pro tier). The AI cost is roughly 6% of subscription revenue per user — healthy margin.
- Development time for the AI layer: approximately 10 weeks, including prompt iteration and cost controls.
- GDPR approach: explicit disclosure in privacy policy, Anthropic's zero-retention API, data minimization per request.
The hybrid approach was not the first plan. The initial prototype used a cloud model for everything, and early testing showed per-user costs that would have made the subscription price untenable. Moving classification and tagging on-device was the single decision that made the business model work.
Your decision checklist
Before you allocate budget for AI integration, work through these questions with your team:
- What specific user problem does AI solve? If you cannot name it in one sentence, you are not ready.
- Can it be done on-device? If yes, start there. It is cheaper, faster, and simpler for privacy.
- What is your per-user API cost ceiling? Model it against your revenue per user. If AI costs exceed 15 to 20% of per-user revenue, the economics are fragile.
- Have you budgeted for cost controls? Rate limits, spending caps, and usage monitoring are not optional — they are launch requirements.
- Is your privacy story clear? Can you explain in plain language what data goes where and why? If not, your DPO and your users will both have questions.
- Do you have the right developer? AI integration is not a junior task. You need someone who understands both iOS architecture and AI provider APIs, including their failure modes and cost structure.
If you have clear answers to all six, you are in a strong position to move forward. If not, the time spent clarifying these will save you far more than it costs.
Considering AI for your iOS app?
We have shipped production AI features on iOS — from on-device classification to cloud-based reasoning — and we understand the trade-offs from both a technical and business perspective. If you want a clear-eyed assessment of what AI would cost and deliver for your specific product, let's talk.
Discuss Your AI Integration ProjectNext in this series
Read next: Hiring a Remote iOS Developer in Europe: What You Should Know — what to look for, how to structure the engagement, and the real cost of hiring iOS talent across European markets.
For developers: the technical details
If you are an engineer evaluating these approaches at the implementation level, our technical series covers the specifics:
- Local-first iOS with SwiftData and CloudKit — schema migration, conflict resolution, and sync in production
- Using Claude API in a Swift App — model routing, streaming, and per-user cost control
- Zero-SDK iOS Analytics — shipping without Firebase, Mixpanel, or Sentry