
I joined a discovery call with a simple brief: "This is basically a copy of Project X. Same timeline."
Project X was a marketing chatbot. Conversational, no proprietary knowledge base, just search integration and personality. We knew that scope well.
Thirty minutes into the call, it's clear this isn't RAG at all. Data processing from S3 buckets, Lambda triggers, ETL pipelines. Table stakes. The real challenge is teaching the model to query and reason over structured data, which makes it a completely different project with a completely different architecture. "Same timeline" for something that shares almost nothing with the original brief.
This happens constantly, and not because clients mislead us. The gap between "AI chatbot" in their head and what that actually requires in practice is just massive. The pre-engineering work, data, access, scoping, validation, is where projects are won or lost, and almost everyone walks in underestimating it.
A client built a Custom GPT over a weekend. They uploaded some PDFs, asked it questions, it worked, and they showed their CEO. Everyone got excited. "We want this, but for the whole company."
That's where it stops being simple. "For the whole company" means multi-tenancy, where different departments see different data. It means role-based retrieval: sales can't access HR documents, legal can't see engineering specs. It means audit logs, access controls, and compliance requirements. Custom GPT handles none of that. It's one user, one knowledge base, no permissions. Scaling from "it works for me" to "it works for the organization" requires a completely different architecture, and the gap between those two things is where most projects get into trouble.
Tools like NotebookLLM and Custom GPTs create a dangerous illusion by hiding all that enterprise complexity. The prototype took a weekend. The production system takes months.
Every client says they have data, but they mean very different things.
Data and credentials need to arrive on day one, and they rarely do. We've waited weeks for database access and months for IT security approvals. One project stalled entirely because a single stakeholder controlled the API credentials and went on vacation. Every week of waiting is a week of zero progress, but in the client's mind the timeline keeps running from the day they signed the contract. The access problem is organizational, and organizations move slowly.
We can usually predict project outcomes from the first call.
Clients who know their bottleneck come in with something specific: "We spend 40 hours weekly on this process. Here are the inputs and outputs. Here's the domain expert who'll validate results." These projects ship because the scope is clear, the outcome is measurable, and there's someone internal who can actually evaluate accuracy.
Clients who want AI everywhere come in with something vaguer: "We want to optimize our processes. We're not sure which ones yet." These projects stall because you can't optimize processes that aren't documented, measure improvement without baselines, or validate AI outputs without domain expertise. Organizational readiness determines the outcome, and these clients haven't gotten there yet.
Successful projects require real, ongoing work from the client side that most teams don't budget for going in.
We build the system, but we can't tell you whether the output is correct for your industry, your regulations, or your edge cases. That judgment has to come from inside.
Before we write code, we need examples of what good looks like: "When users ask X, good answers look like Y." Hundreds of them. Without this, there's no way to measure whether we're making real progress or just producing wrong answers with more confidence.
Getting to 85% accuracy might take 6 weeks. Getting to 95% might take another 6. Getting to 99% might be impossible with the data quality available, and those last few percentage points, serving maybe 2% of edge cases, could consume 40% of the remaining budget. That's a business decision, and it belongs to the client.
When source documents change, someone has to update them. When accuracy drifts, someone has to investigate. This is an ongoing operation that needs ownership, and teams that treat it as a one-time build tend to find out the hard way.
Most clients expect to hand off requirements and receive a finished product. AI development requires continuous involvement from the people who understand the domain, and projects that don't have that commitment built in rarely make it to production.
Best case scenario: clean data, clear scope, an engaged stakeholder with domain knowledge. Six to eight weeks, with most of that time going to prompt engineering and iteration rather than infrastructure.
But "clean data" is rare, "clear scope" requires serious upfront work, and "engaged stakeholder" means someone's calendar is genuinely blocked for this project rather than squeezed between other priorities. When any of these conditions are missing, the timeline multiplies. When all three are missing, it's worth reconsidering whether to start at all.
Projects rarely fail for technical reasons. The failures are almost always organizational.
We deliver a working system and it sits in staging because the client doesn't have engineering resources to integrate it. They budgeted for building the thing, not for deploying it.
Midway through, the client realizes the problem they originally described isn't actually their biggest pain point. The AI works fine; the business case behind it didn't hold up.
We lay out the math, that last 5% of accuracy to cover edge cases will cost 40% of the remaining budget, and the client wants it anyway. Then the budget runs out and the project gets labeled "over scope." None of these are engineering problems.
Before signing contracts, dig into the actual data. Not descriptions of it, the data itself. We run a Rapid Validation Sprint: four weeks of real data access, real complexity mapping, and real unknowns surfaced. Only then do we estimate, based on what we actually found rather than what we assumed going in.
The companies quoting 50% less aren't doing this work. They're estimating blind, and when the data turns out messier than expected, and it always does, they either blow the budget or cut scope.
RAG tutorials make this look straightforward: upload documents, chunk them, embed them, query them. Production is a different story. Data is messy, access is slow, validation requires domain expertise you don't have, and accuracy expectations routinely exceed what the underlying data can actually support.
The engineering is the manageable part. Data quality, access, organizational alignment, and validation are where projects actually succeed or fail, and those are problems that show up long before anyone writes a line of code.
Most AI initiatives struggle because the organization isn't ready. Data isn't organized. Processes aren't documented. Nobody's been assigned to validate outputs. That's not a criticism, it's just the reality most teams are starting from.
The question isn't whether AI can help your business. It's whether your business is ready to help the AI.
I joined a discovery call with a simple brief: "This is basically a copy of Project X. Same timeline."
Project X was a marketing chatbot. Conversational, no proprietary knowledge base, just search integration and personality. We knew that scope well.
Thirty minutes into the call, it's clear this isn't RAG at all. Data processing from S3 buckets, Lambda triggers, ETL pipelines. Table stakes. The real challenge is teaching the model to query and reason over structured data, which makes it a completely different project with a completely different architecture. "Same timeline" for something that shares almost nothing with the original brief.
This happens constantly, and not because clients mislead us. The gap between "AI chatbot" in their head and what that actually requires in practice is just massive. The pre-engineering work, data, access, scoping, validation, is where projects are won or lost, and almost everyone walks in underestimating it.
A client built a Custom GPT over a weekend. They uploaded some PDFs, asked it questions, it worked, and they showed their CEO. Everyone got excited. "We want this, but for the whole company."
That's where it stops being simple. "For the whole company" means multi-tenancy, where different departments see different data. It means role-based retrieval: sales can't access HR documents, legal can't see engineering specs. It means audit logs, access controls, and compliance requirements. Custom GPT handles none of that. It's one user, one knowledge base, no permissions. Scaling from "it works for me" to "it works for the organization" requires a completely different architecture, and the gap between those two things is where most projects get into trouble.
Tools like NotebookLLM and Custom GPTs create a dangerous illusion by hiding all that enterprise complexity. The prototype took a weekend. The production system takes months.
Every client says they have data, but they mean very different things.
Data and credentials need to arrive on day one, and they rarely do. We've waited weeks for database access and months for IT security approvals. One project stalled entirely because a single stakeholder controlled the API credentials and went on vacation. Every week of waiting is a week of zero progress, but in the client's mind the timeline keeps running from the day they signed the contract. The access problem is organizational, and organizations move slowly.
We can usually predict project outcomes from the first call.
Clients who know their bottleneck come in with something specific: "We spend 40 hours weekly on this process. Here are the inputs and outputs. Here's the domain expert who'll validate results." These projects ship because the scope is clear, the outcome is measurable, and there's someone internal who can actually evaluate accuracy.
Clients who want AI everywhere come in with something vaguer: "We want to optimize our processes. We're not sure which ones yet." These projects stall because you can't optimize processes that aren't documented, measure improvement without baselines, or validate AI outputs without domain expertise. Organizational readiness determines the outcome, and these clients haven't gotten there yet.
Successful projects require real, ongoing work from the client side that most teams don't budget for going in.
We build the system, but we can't tell you whether the output is correct for your industry, your regulations, or your edge cases. That judgment has to come from inside.
Before we write code, we need examples of what good looks like: "When users ask X, good answers look like Y." Hundreds of them. Without this, there's no way to measure whether we're making real progress or just producing wrong answers with more confidence.
Getting to 85% accuracy might take 6 weeks. Getting to 95% might take another 6. Getting to 99% might be impossible with the data quality available, and those last few percentage points, serving maybe 2% of edge cases, could consume 40% of the remaining budget. That's a business decision, and it belongs to the client.
When source documents change, someone has to update them. When accuracy drifts, someone has to investigate. This is an ongoing operation that needs ownership, and teams that treat it as a one-time build tend to find out the hard way.
Most clients expect to hand off requirements and receive a finished product. AI development requires continuous involvement from the people who understand the domain, and projects that don't have that commitment built in rarely make it to production.
Best case scenario: clean data, clear scope, an engaged stakeholder with domain knowledge. Six to eight weeks, with most of that time going to prompt engineering and iteration rather than infrastructure.
But "clean data" is rare, "clear scope" requires serious upfront work, and "engaged stakeholder" means someone's calendar is genuinely blocked for this project rather than squeezed between other priorities. When any of these conditions are missing, the timeline multiplies. When all three are missing, it's worth reconsidering whether to start at all.
Projects rarely fail for technical reasons. The failures are almost always organizational.
We deliver a working system and it sits in staging because the client doesn't have engineering resources to integrate it. They budgeted for building the thing, not for deploying it.
Midway through, the client realizes the problem they originally described isn't actually their biggest pain point. The AI works fine; the business case behind it didn't hold up.
We lay out the math, that last 5% of accuracy to cover edge cases will cost 40% of the remaining budget, and the client wants it anyway. Then the budget runs out and the project gets labeled "over scope." None of these are engineering problems.
Before signing contracts, dig into the actual data. Not descriptions of it, the data itself. We run a Rapid Validation Sprint: four weeks of real data access, real complexity mapping, and real unknowns surfaced. Only then do we estimate, based on what we actually found rather than what we assumed going in.
The companies quoting 50% less aren't doing this work. They're estimating blind, and when the data turns out messier than expected, and it always does, they either blow the budget or cut scope.
RAG tutorials make this look straightforward: upload documents, chunk them, embed them, query them. Production is a different story. Data is messy, access is slow, validation requires domain expertise you don't have, and accuracy expectations routinely exceed what the underlying data can actually support.
The engineering is the manageable part. Data quality, access, organizational alignment, and validation are where projects actually succeed or fail, and those are problems that show up long before anyone writes a line of code.
Most AI initiatives struggle because the organization isn't ready. Data isn't organized. Processes aren't documented. Nobody's been assigned to validate outputs. That's not a criticism, it's just the reality most teams are starting from.
The question isn't whether AI can help your business. It's whether your business is ready to help the AI.
