Before 2023, there was a clear line between “automatable” and “not automatable.” If a task followed a predictable pattern with structured data, you could automate it. If it required judgment, interpretation, or contextual understanding, you could not. The line was clear, and it had not moved much in 20 years.
LLMs erased that line. Not by automating everything — but by making a specific category of tasks automatable that genuinely were not before.
What AI is actually good at today
Classification and routing. “Is this email a complaint, a question, or a feature request?” An LLM can classify incoming communication with high accuracy and route it to the right person or queue.
AI email classification achieves 98% accuracy, reducing manual sorting by 85–95% and saving roughly 4 minutes per processed email. This replaces the human who reads every email and forwards it manually.
Data extraction from unstructured sources. Pulling structured information from PDFs, emails, screenshots, or free-text descriptions. “Extract the invoice number, date, amount, and vendor from this PDF” is a task that modern LLM pipelines handle with 95%+ accuracy — even across different formats and languages.
This is the capability that eliminated the biggest category of “unautomatable” manual work.
Contextual text generation. Writing responses that need to be appropriate to the context — review responses, follow-up emails, status summaries. Not creative writing. Functional writing where the structure is predictable but the content varies. The Guest Suite story is the proof.
Summarization and synthesis. Condensing long documents, meeting transcripts, or comment threads into actionable summaries. Studies show a 1.47% hallucination rate in medical note summarization — good enough for drafts, but review remains essential for final documents.
The “LLM as a step” pattern
All major automation platforms now support LLM integration as a workflow step. The pattern is straightforward: Trigger → Fetch data → LLM classifies/extracts/generates → Route to action → Complete.
You can use this pattern for email triage, lead scoring, content moderation, data enrichment, and invoice processing.
Both
Zapier and
Make offer built-in LLM steps, and
n8n provides full control over model selection and prompting. This is the lowest-friction way to add AI to your existing workflows — no custom code required.
What AI is not good at (yet)
Precision-critical calculations. Do not use an LLM to compute your quarterly revenue. Use a spreadsheet or a database query. LLMs are language tools, not calculators.
Tasks requiring perfect reliability. LLMs are probabilistic. They will occasionally produce wrong outputs that look right. For tasks where a 1% error rate is unacceptable (financial reporting, compliance documents, medical records), AI should assist, not replace.
Tasks requiring access to real-time internal data. Unless you build a retrieval-augmented generation (RAG) pipeline or give the model access to your systems, it cannot answer questions about your specific data. “How many orders did we ship last week?” is a database query, not an AI task.
Full autonomy. Agentic AI — fully autonomous agents that coordinate across multiple systems — is emerging but still early.
Gartner predicts over 40% of agentic AI projects will be canceled by 2027. MIT’s NANDA study found that 95% of enterprise GenAI pilots deliver zero measurable return. The 5% that succeed start small and prove value before scaling.
Start with LLM-as-a-step, not full autonomy.
What does AI automation actually cost?
The cost conversation around AI is either “it is basically free” or “it will bankrupt you.” Neither is true. Here are real numbers:
- Simple classification: ~€0.30 per 1,000 tasks (using a lightweight model)
- Email response generation: ~€2 per 1,000 tasks (mid-tier model)
- Complex document analysis: ~€90 per 1,000 tasks (premium model)
At roughly €50/month in LLM costs, you can automate work that would take a human 5–10 hours — easily €200+ in labor savings.
Cost optimization makes it even cheaper: route 70% of routine tasks to lightweight models and only 30% of complex cases to premium ones. Caching repeated queries cuts costs by 10×. Batch APIs offer 50% savings on non-time-sensitive work.
Builder’s note
The pattern I recommend for AI-augmented automation is always human-in-the-loop, at least initially. The AI does the heavy lifting — drafting, classifying, extracting — and a human reviews the output before it goes live. This gives you 90% of the time savings while keeping the error rate near zero. As confidence grows, you can gradually reduce human oversight on the cases where the AI consistently gets it right.
The browser extension pattern
One automation pattern I find particularly effective is the AI-powered browser extension. Instead of building a separate application and asking your team to change their workflow, you bring the automation to where they already work.
The Guest Suite review response tool is a good example of this pattern. The team was already on the review platforms. They did not need a new dashboard, a new login, a new workflow. They needed the same workflow, but with the hard part handled.
The browser extension sits in the background, activates when you view a review, and offers a generated response. Accept it, edit it, or ignore it. No workflow change required.
This pattern works because it respects how people actually work. The best automation is invisible — it removes friction without adding complexity.
And it is getting easier.
Anthropic’s Claude now runs natively in Chrome, summarizing pages, drafting responses, and extracting data directly from whatever tab you are viewing.
Claude Cowork takes this further — the AI watches your screen, understands what you are doing, and proactively suggests actions.
On the no-code side, AxAxiom.ai lets you build browser automations visually: record a workflow of clicks and inputs, and it replays them on schedule.
These are not “tools that happen to work in a browser.” They are automation that meets people where they already spend their day.