Skip to content

I fix AI adoption, train developers, and build the rest.

Most AI rollouts stall at the human layer, not the tech. I come in, find what's blocking adoption, set up the workflows and tooling that make it stick, train your developers to run them well, and build custom systems when nothing off the shelf will do.

How I can help

Consult, train, or build — depending on what you need.

AI tools were rolled out. Adoption stalled.

Typical AI tool adoption ~20%
With structured adoption ~80%

Implementation consulting

I come into your company, diagnose why AI adoption isn't working, and build the setup that fits your stack — together with you. Workflows, tooling configuration, and the team rituals that keep working long after I leave.

On-site diagnosis of where adoption breaks
Custom tool configuration built around your stack
Pair sessions until the new workflows stick
Measurable productivity shift, not vibes
Learn more

Your developers need to work with AI, not just hear about it.

Off-the-shelf workshop
Same slides for everyone
Generic demos
Fixed format
Tailored to you
Your tech stack
Your team's level
Your format

AI & developer training

Educational workshops in whatever format fits — half-day deep dives, multi-week programmes, online or on-site. Hands-on with Claude Code, Cursor, GitHub Copilot on real codebases. Commissioned by Cegos Integrata, IT-Schulungen.com, and other L&D providers.

AI-assisted coding with Claude Code, Cursor & more
Hands-on workshops with real codebases
Prompt engineering and workflow design
Online, on-site, or hybrid delivery
Learn more

You need it built — not just spec'd.

Problem
Architecture
Production

Custom development

Complete products from concept to production. AI systems, real-time applications, and full-stack web apps — business problem first, then code.

AI integration (OpenAI, Anthropic, custom models)
Real-time features with Phoenix LiveView
RAG and semantic search (Qdrant, pgvector)
Full-stack web applications (Elixir/Phoenix, PostgreSQL)
Learn more

Why I deliver results

Strategic depth meets technical precision

Most AI initiatives stall at the human layer, not the technology. My background sits at that intersection — understanding how teams actually adopt new tools, and the engineering it takes to build them. That's why the systems I deliver get used, not just shipped.

Psychology BSc

How people actually adopt new tools

Trained in how teams actually adopt — or quietly reject — new tools. The reason adoption stalls is rarely technical, and treating it as such is the most common mistake.

Management MA

Business and engineering, both fluent

Fluent in both business goals and engineering trade-offs. I translate between CFO concerns and architecture decisions, so nothing important gets lost in handoff.

Co-Founder & CTO, The Happy Beavers

6 years scaling systems and teams

Six years scaling systems and leading teams at The Happy Beavers — automating 80% of internal operations and shipping production software end to end.

SDW scholar

Top 2% nationally, three years running

Triple-recipient of the Stiftung der Deutschen Wirtschaft scholarship — awarded to the top 2% of students nationally for academic and leadership excellence.

Luka Breitig

About me

Technical Product Builder

I'm Luka Breitig, a Technical Product Builder who bridges business strategy and technical execution. Dual background in Psychology and Management combined with hands-on engineering. I build complete products, help teams adopt AI workflows, and train developers — always starting with the business problem, not the tech stack. Three production systems live and serving real users.

"Technology only works when people trust it. Whether I'm building a product, consulting on adoption, or training a team — I start with understanding the people, not just the code."

— My philosophy
Top 2%
SDW scholar
Production
Systems built

Blog

Practical insights on AI technologies and real-world applications

Do LLMs really write better Elixir than Python?

Do LLMs really write better Elixir than Python?

April 02, 2026
Luka Breitig

A Tencent benchmark ranked Elixir #1 for LLM code generation. As an Elixir developer, I dug into the data — the truth is more nuanced than the headline, but the structural advantages are real.

From chatbot to agent: a step-by-step mental model for understanding agentic AI

From chatbot to agent: a step-by-step mental model for understanding agentic AI

January 04, 2026
Luka Breitig

Moving from conversational mimicry to autonomous execution: understand how agentic AI systems pursue complex goals, use tools, and adapt to changing environments through a continuous reasoning loop.

Insights from the field and production-grade AI development

To the blog

Frequently asked questions

Common questions about AI adoption, consulting engagements, and how the work fits together

A typical engagement runs in three phases. First I embed with your team — observing real workflows, current tool usage, and where the friction actually sits. Then I design and configure the setup: agent and tooling configuration, prompts and playbooks, and the integrations that make the workflow practical for your stack. Finally, I work alongside your developers in pair sessions, code reviews, and iteration cycles until the new practices stick. Scope is agreed per outcome, not per hour.

We agree on baseline metrics before the engagement starts — typically PR cycle time, time-to-first-commit on new tasks, share of pull requests touched by AI tools, and developer satisfaction. After the workflows are embedded, we re-measure. The goal is a defendable shift you can show your CFO, not a vibes-based 'it feels faster'. If the numbers do not move, we have a problem to solve together rather than a result to celebrate.

Training is educational: workshops, hands-on sessions, and multi-day programmes that teach your developers how to use AI coding tools effectively. It is the right fit when the gap is primarily knowledge. Consulting goes further: I diagnose why adoption is not sticking, design the workflow and tooling setup, and embed it with your team over several weeks. Often the answer is both — first set up the system, then train the wider organisation on how to use it well.

It is almost never the technology. The recurring patterns are tools rolled out without configuration for the actual codebase, no clear guidance on when to use AI versus write by hand, no shared prompts or playbooks, and no follow-through after the initial training session. My psychology background combined with hands-on engineering means I work on both layers — the human one and the technical one — and diagnose the real bottleneck before suggesting solutions.

That is the whole point. I leave behind documented playbooks, configured tooling (CLAUDE.md, agent configurations, prompt libraries), and at least one internal champion who can extend the setup. The success criterion is not 'Luka is still in the building' — it is that productivity gains compound after I am gone. Top-up training and ad-hoc consulting are available when something genuinely new comes up, but they are optional, not built-in dependencies.

Off-the-shelf tools are excellent for general tasks. Custom AI matters when you need deep integration with your stack, use of proprietary data, or enforcement of complex business rules. The systems I build automate multi-step workflows — syncing CRM data, running inventory checks, routing approvals — which a chatbot or simple API wrapper cannot do. The decision usually comes down to whether the workflow is your competitive edge or just a productivity helper.

Have a specific question about your AI project?