ai
AI for Internal Tools
Looking for ai for internal tools? I'm Olle Evertsson, an independent fullstack engineer based in Stockholm helping teams ship production-grade systems across ai assistant company, internal ai assistant, ai for teams.
Stockholm is full of developers — but few have deep experience shipping production-grade systems end-to-end. My focus is work that actually ships, gets paid for by real customers, and is maintained over time.
In practice I work with: AI assistant company, internal AI assistant, AI for teams.
Why work with me
- Senior fullstack experience across Next.js, React Native, AI, and Postgres.
- Fixed price per phase — no hourly billing for uncertain work.
- You own 100% of the code, infrastructure, and credentials.
- Delivered from Stockholm, working with clients across the Nordics and EU.
- GDPR-compliant architecture from sprint one — not bolted on afterwards.
How we work together
01
Discovery
We map the problem, your existing stack, and business goals on a 30-min call.
02
Proposal
Within 48 hours you get a concrete proposal with scope, milestones, and fixed price.
03
Build
Weekly demos, full repo access, and continuous deploys to staging.
04
Launch & support
We launch together and you can choose a support retainer for continued work.
Frequently asked questions
Which AI models do you use?+
Claude Opus 4.6 and Sonnet 4.6 (Anthropic) as defaults for reasoning and text, GPT-5 for specific use cases, Gemini 2.5 Pro for Google Cloud stacks, and open source (Llama 3.3, Mistral Large) when data must stay on-prem. Model choice is always driven by the problem, not the hype.
How do you prevent hallucinations?+
Retrieval Augmented Generation (RAG) against your sources, strict prompt design, Zod-validated output schemas, and guardrails. Critical use cases get human-in-the-loop for the last 5%.
What does running AI in production cost?+
Token costs for Claude/GPT typically run €50–€500/month for a B2B app with a few hundred active users. We optimize via caching, model routing, and prompt compression.
How long for an AI integration?+
PoC in 2–4 weeks. Production-grade AI feature with eval pipeline, monitoring, and guardrails: 6–10 weeks.
Is AI safe for sensitive enterprise data?+
Yes, when implemented correctly. Anthropic, OpenAI, and Google do not train on API data. For extra-sensitive workloads we run models on-prem or via Azure OpenAI/AWS Bedrock with EU data zoning.
Related services
Ready to start?
Book a 30-min strategy call