Loading...
Loading...
Weekly AI insights —
Real strategies, no fluff. Unsubscribe anytime.
Founder & CEO, Agentik {OS}
Make transforms data between every step, not just connects apps. Add AI routing and classification and you get workflows that actually think.

Make has a reputation problem. Most people discover it as a cheaper Zapier alternative. They build a few simple three-step automations, appreciate the visual interface, and leave it at that. They miss what actually makes Make powerful.
Make is a data transformation platform. Every module in a scenario can reshape, filter, and manipulate data before passing it to the next step. Add AI classification to that data transformation capability, and you can build workflows that handle extraordinary complexity without any custom code.
I have been running Make scenarios in production for two years. Here is what actually matters.
This comparison comes up constantly. My take after running both in production:
| Dimension | Make | n8n |
|---|---|---|
| Visual design | Excellent | Good |
| Data transformation | Outstanding | Good |
| Self-hosting | No | Yes |
| Pricing model | Operations/month | Self-hosted = fixed |
| Setup complexity | Low | Medium-High |
| AI integration | Via HTTP/OpenAI module | Native AI nodes |
| Error handling | Strong | Strong |
| Best for | Complex data transforms, team use | High volume, full control |
Make wins on data transformation and ease of use. n8n wins on pricing at scale and infrastructure control. Most serious automation builders end up using both for different purposes.
Make's scenario builder makes complex data manipulation visual in a way that n8n's flow-based approach cannot match. When your workflow involves reshaping arrays, aggregating data across sources, or complex conditional mapping, Make's interface is genuinely superior.
For teams without DevOps capability who still need serious automation, Make is the right choice despite the per-operation pricing. The operational overhead of self-hosting n8n is real.
Make's most underused feature is its built-in data transformation functions. Every connection between modules can run transformations without a code module.
String functions: toString, trim, replace, substring, split, join. Array functions: map, filter, sort, slice, flatten. Math functions: round, floor, ceil, min, max. Date functions: formatDate, parseDate, addDays, dateDifference.
And crucially: these can be chained and nested.
// Example: Clean and format a customer name from a messy API response
// Raw: " JOHN SMITH "
// Target: "John Smith"
trim(toString(join(map(split(lowercase(trim(raw_name)), " "),
(word) => capitalize(word)), " ")))
// Reads inside-out:
// 1. trim whitespace
// 2. lowercase everything
// 3. split on spaces
// 4. capitalize each word
// 5. rejoin with spaces
// 6. ensure string type
// 7. trim again
This kind of data normalization used to require a code module or a custom function. In Make, it is inline transformation at the module connection level.
The pattern that transforms Make from automation tool to intelligent system: use an OpenAI or HTTP module to call an AI API, then route based on the structured response.
My standard approach uses the HTTP module rather than the native OpenAI module. More control, better error visibility.
// HTTP Module Configuration
// URL: https://api.anthropic.com/v1/messages
// Method: POST
// Headers: x-api-key: {{your_key}}, anthropic-version: 2023-06-01
// Body:
{
"model": "claude-haiku-20241022",
"max_tokens": 256,
"messages": [{
"role": "user",
"content": "Classify this support ticket. Return ONLY valid JSON: {\"category\": \"billing|technical|account|other\", \"priority\": \"low|medium|high\"} Ticket: {{1.ticket_body}}"
}]
}
// Parse the response content as JSON
// Map to downstream router moduleThe key: ask for ONLY JSON with no surrounding text. AI models will often wrap JSON in markdown code fences or add explanatory text unless you explicitly instruct otherwise. Short, direct prompts that return structured data work better than elaborate prompts in this context.
Make's Router module splits execution into multiple paths based on conditions. After the AI classification module:
{{2.category}} = "billing" and {{2.priority}} = "high" goes to billing team Slack with urgent flag{{2.category}} = "technical" goes to engineering Jira with full context{{2.category}} = "account" goes to account team with CRM lookupEach route can have completely different downstream logic. The AI classification is just the decision input. Make handles everything that follows.
A new lead enters your CRM. The scenario:
The AI scoring module does something rules cannot: it handles novel combinations of signals. A small company with a sophisticated tech stack scores higher than a large company still on legacy systems. Rules cannot capture that nuance. AI can.
Every piece of content submitted by contractors passes through an AI quality gate before publication.
// AI Quality Assessment Prompt
const qualityPrompt = `
Evaluate this content submission on these dimensions.
Return JSON only:
{
"passes": true/false,
"score": 0-100,
"issues": ["specific issue 1", "specific issue 2"],
"strengths": ["strength 1", "strength 2"],
"revision_priority": "none|minor|major"
}
Evaluation criteria:
- Factual accuracy claims (flag unverified statistics)
- Tone consistency with brand voice
- Required sections present
- Minimum word count met
- No placeholder text remaining
Content: {{1.submission_text}}
`;Submissions that fail go back to the contractor with specific feedback. Submissions that pass go to human editor review for final approval. The AI handles the mechanical quality checks. The human handles judgment calls.
Result: editor time on each piece drops from 45 minutes to under 10, because the mechanical feedback cycle is eliminated.
Not every notification deserves a Slack ping. Some should be email. Some should be SMS. Some should wait for business hours.
AI module takes the event data and outputs:
This eliminates notification fatigue. High-signal events cut through. Low-signal events wait.
Incoming invoices via email attachment. The scenario:
This scenario eliminates manual data entry entirely for routine invoices. The AI handles the messy OCR output that rule-based parsers choke on.
Run nightly. Pull all customer activity data: logins, feature usage, support tickets, payment history, NPS scores. Feed to AI for health assessment. Flag accounts showing early churn signals for CSM follow-up.
The AI notice patterns rules miss: a customer who logs in daily but only uses one feature despite having access to five is showing a risk signal that pure activity metrics hide.
Make's biggest strength and most common friction point is arrays. Understanding how Make handles arrays unlocks serious power.
The Iterator module splits an array into individual items for processing. The Aggregator module collects processed items back into an array. Most complex Make scenarios use this pattern:
Get array of items from API
-> Iterator (process each item)
-> Transform/enrich each item
-> Apply AI logic to each item
-> Aggregator (collect results)
-> Format consolidated output
-> Send report or store data
The mistake beginners make: trying to process arrays without iterating. Make processes one bundle at a time. If your data is an array, you must iterate.
Many APIs return paginated results. Make's repeater module handles this:
This pattern works for any API that uses offset/limit or page/per-page pagination.
Make's error handling is thorough but requires intentional setup.
Every module has an error handler attachment point. Click the wrench icon, add an error handler route. For critical modules, I always add error handling that:
The scenario-level error handling (Operations > Add error handler) covers uncaught errors that bubble past module-level handlers. Set this up on every production scenario.
| Error Type | Handling Strategy |
|---|---|
| Rate limit (429) | Retry with exponential backoff |
| Network timeout | Retry up to 3 times |
| Invalid data structure | Log and skip, alert team |
| Authentication failure | Alert immediately, stop scenario |
| API down | Queue for retry when restored |
Make works best as part of a larger system, not as a monolith.
For workflows requiring custom logic that Make's built-in tools cannot handle cleanly, use the HTTP module to call your own API. A simple serverless function handles the complex logic and returns a clean result to Make. This gives you the best of both worlds: Make's excellent orchestration and your code's flexibility.
For high-volume AI operations, consider batching requests through the HTTP module rather than one call per item. Claude and OpenAI APIs respond faster to well-structured batch prompts than to many small requests.
For monitoring, Make's built-in execution history is good but not sufficient for production. Export execution data to a proper database or analytics tool. You want to track scenario success rates over time, not just look at individual executions.
Pair Make with n8n for self-hosted high-volume workflows and you cover the full spectrum of automation needs without compromising on either capability or cost.
Make's pricing is operation-based. Each module execution costs one operation. A scenario with ten modules running a thousand times per month costs ten thousand operations.
The Core plan includes ten thousand operations per month. The Pro plan includes a hundred thousand. Beyond that, you are buying additional operation packs.
For scenarios that run rarely or with low module counts, Make is very affordable. For scenarios that run continuously with many modules and AI steps, the costs can surprise you. Calculate your expected monthly operations before committing to a complex scenario.
AI API calls through the HTTP module count as one Make operation each, but the AI API cost itself is separate. Both need to be in your budget.
Make is the right automation platform when:
When to choose something else: extremely high volume (consider n8n self-hosted), workflows that need sub-second response times (consider code-first solutions), workflows that are essentially just code with triggers (just write the code).
For most businesses building their first intelligent automation stack, Make is the fastest path from idea to production workflow. The visual interface, the rich integration library, and the data transformation capabilities make it genuinely powerful without the operational complexity of self-hosted alternatives.
Q: What is Make (formerly Integromat)?
Make is a visual automation platform that connects apps and services through workflows called scenarios. It uses a visual editor where you drag and connect modules representing different services. Make excels at complex integrations with its intuitive visual interface and robust error handling.
Q: How do you build AI workflows in Make?
Build AI workflows in Make by combining HTTP modules (calling AI APIs like Claude or GPT), data transformation modules, and app-specific modules (CRM, email, databases). Make handles the orchestration, error handling, and scheduling while you design the logic visually.
Q: What are the best Make + AI integration patterns?
Best patterns include AI-powered email processing (classify and route automatically), content generation pipelines (ideation to publish), data enrichment (AI analyzing and categorizing raw data), customer communication (personalized responses at scale), and document processing (extract, analyze, and file).
Full-stack developer and AI architect with years of experience shipping production applications across SaaS, mobile, and enterprise. Gareth built Agentik {OS} to prove that one person with the right AI system can outperform an entire traditional development team. He has personally architected and shipped 7+ production applications using AI-first workflows.

n8n Workflow Automation: Build AI-Powered Decision Engines
n8n connects apps with intelligence. Self-hosted, no per-execution pricing, and every step can use AI to analyze, decide, and adapt. Build it right.

Stripe Billing Automation: Get the Scary Parts Right
Billing bugs charge people twice or miss charges entirely. Stripe handles subscriptions, usage metering, dunning, and tax so your revenue doesn't leak.

Prompt Engineering: The Craft Behind Reliable AI Output
The difference between garbage and production-quality AI output is not magic. It is craft. System prompts, few-shot examples, chain-of-thought, structure.
Stop reading about AI and start building with it. Book a free discovery call and see how AI agents can accelerate your business.