Addepto in now part of KMS Technology – read full press release!

in Blog

March 24, 2026

How to Choose the Right Data Company in 2026

Author:




Edwin Lisowski

CGO & Co-Founder


Reading time:




18 minutes


Choosing a data engineering or MLOps partner is one of the more consequential decisions a data team makes — not because it’s technically complex, but because the wrong choice compounds. Bad pipelines generate bad data. Bad data generates bad models. Bad models generate bad decisions. And unwinding all of it costs far more than getting the selection right the first time.

This guide gives you a framework for doing that: a way to diagnose where you are, what to look for, and how to run a process that produces real evidence rather than polished proposals.

Key takeaways

  • Most data partnerships fail not because of technology, but because of a mismatch between what a vendor can actually deliver and what they led you to believe during the sale.
  • Where you are starting from — first platform, existing infrastructure, or AI in production — should determine how you evaluate and what you prioritise. A framework that treats all three the same will mislead you.
  • The clearest signal of a partner who thinks strategically is what they ask you, not what they show you.
  • MLOps is not something you can add later. By the time you realise you need it, you are already paying for not having it.
  • A well-designed pilot is the single most reliable risk-management tool available. What matters is not just whether it works — it is what the vendor finds along the way.
  • The commercial structure of a partnership tells you something about alignment. So do the exit terms — and most buyers only read them after they wish they had read them sooner.

Data Engineering Service - CTA

In November 2012, National Grid US went live with a new SAP platform five days after Hurricane Sandy struck the northeast — because delaying would have required government approval for a rate increase. The system failed immediately. Payroll calculations were wrong. Supply chain processes stopped.

Within months, stabilisation efforts were running at $30 million a month. The cleanup took two years and cost $585 million — 150% of the original implementation budget. National Grid later sued its systems integrator, alleging it had misrepresented its capabilities and assigned consultants with, in the words of the court filing, “virtually no experience implementing an SAP platform for a US-regulated utility.” Wipro settled for $75 million without admitting fault.

National Grid is an extreme case, but the dynamic is not unusual. BCG’s research across more than 1,000 companies found that two-thirds of large-scale technology programmes fail to deliver on time, within budget, or to defined scope. The consistent root cause is not the technology — it is partner selection risk: a mismatch between a vendor’s actual capabilities and what they represented during the sale.

This guide provides a reusable framework for avoiding that mismatch: three buyer situations, seven evaluation criteria, a scoring rubric, and a pilot design, so the choice can be made on evidence rather than sales theatre.

1. Start with your situation

The right partner depends heavily on where an organisation is starting from. A company building its first data warehouse has fundamentally different needs from one migrating a Teradata estate or one trying to get ML models into production governance. Three starting situations cover most mid-market and enterprise buying scenarios.

Situation A — Greenfield: building the first real data platform

Data sits in SaaS tools — Salesforce, HubSpot, Shopify, Google Analytics. Analysts pull exports manually. There is no single source of truth for revenue, product, or customer metrics. Engineering is occupied with the product roadmap, so data requests accumulate.

What this situation calls for is design authority: a partner opinionated enough to make architectural decisions the internal team is not yet equipped to make, build the first ELT pipeline and warehouse layer, and hand it over in a maintainable state within six to twelve months.

Watch out for:
Over-engineered solutions. A sophisticated lakehouse architecture is the wrong answer for a team of three analysts. If a vendor’s first proposal involves streaming infrastructure and a federated data mesh, ask whether they have read the brief or their own playbook.

Situation B — Scaling: modernising existing infrastructure

Pipelines, a warehouse, and BI dashboards exist — but all of it is starting to creak. Cloud costs are climbing without clear attribution. A migration project has stalled because no one documented the legacy dependencies. Analysts are waiting days for data that should take hours.

This situation calls for a partner with a proven track record at comparable data volumes and architecture complexity. FinOps awareness is non-negotiable. A partner who does not factor cost design into the architecture from day one will deliver a working platform that costs twice what it should to run.

Questions to ask in discovery:

  • “Walk us through a migration at our scale. What failed, what changed, what would you do differently?”
  • “If our data volume doubled in six months, what in your proposed architecture would break first?”
  • “How do you handle schema evolution without manual rework?”

Situation C — AI-curious: stuck between experimentation and production

ML models exist — in Jupyter notebooks. They work in the notebook. In production, they fail or silently degrade. No one knows when predictions have drifted. Retraining is triggered by a complaint, not a threshold. The compliance team is asking how the models make decisions, and there is no satisfying answer.

Enforcement of EU AI Act high-risk AI system requirements starts August 2026, covering fraud detection, credit scoring, and HR-related models. For many organisations in regulated industries, that deadline has already made this situation urgent.

This situation requires not a data engineering generalist but an MLOps partner able to demonstrate a full lifecycle — from raw data to monitored, retrain-capable model in production — with specific examples from comparable regulated environments.

2. Seven criteria for evaluating any partner

The following criteria work as both an evaluation lens and a structured question set. Credible evidence means referenced examples and documented artefacts — not marketing claims.

Criterion 1 — Business and domain alignment

The clearest signal that a partner thinks strategically rather than technically is what they ask in the first conversation. Do they ask about downstream data consumers, how data connects to revenue, and what happens when a pipeline fails? Or do they propose a technology stack before they understand the problem?

According to dbt Labs’ 2025 State of Analytics Engineering report, 56% of data teams cite poor data quality as their biggest challenge, and organisations addressing data quality first show 2.5x higher transformation success rates. A partner who leads with tooling rather than with “what decisions does this data need to enable?” is revealing their orientation.

 

🚩 Red Flag What it signals
Proposes a tool stack in the first meeting, before understanding the business model Tool-first thinking — the partner is selling a solution before diagnosing a problem
Cannot name the downstream business users of the data they would be building No genuine discovery has taken place; the proposal is generic
No questions about regulatory context, data literacy of stakeholders, or revenue model The partner is not treating this as a business engagement

Criterion 2 — Technical stack and patterns fit

Platform expertise matters — but in 2026 the differentiator is governance and AI integration, not raw query performance.

The three dominant platforms have substantially converged on capability.

 

Platform Structural advantage Ask your partner for
Databricks Unity Catalog supports Delta Lake and Apache Iceberg natively; strongest end-to-end MLOps via Mosaic AI Architecture diagrams showing Unity Catalog setup; Databricks Partner Connect status
Snowflake Broadest third-party connector ecosystem; SQL-first; strongest for heavy integration requirements SnowPro certification; specific delivery examples, not just “we know Snowflake”
BigQuery Serverless architecture eliminates disk-spill penalties; best for Google Cloud-native orgs with Vertex AI on the roadmap Vertex AI integration experience; Looker/Dataflow delivery examples

 

Modern stack baseline to verify:

  • Orchestration: Apache Airflow, Prefect, or Dagster — not cron jobs
  • Transformation: dbt (analytics engineering layer)
  • Ingestion: Fivetran, Airbyte, or custom Kafka/Spark pipelines
  • Observability: Monte Carlo, Evidently AI, or Great Expectations
  • Infrastructure: Terraform or Pulumi (IaC), not manual console deployments

 

🚩 Red Flag What it signals
“We’re a Databricks shop” applied to every client regardless of workload Familiarity bias — the partner recommends what they know, not what fits
Cannot explain why they chose specific tools for comparable projects No deliberate architecture reasoning; tools are chosen by default
Data quality and observability absent from the proposal, or positioned as phase 2 Governance is an afterthought; technical debt is being deferred to you

 

A competent partner should walk through their drift detection approach without prompting: establishing baseline distributions at deployment, continuously computing the same statistics over sliding windows on production data, and using statistical tests to compare.

If they cannot, they are operating as a data science consultancy — not an MLOps partner.

EU AI Act — August 2026 enforcement:
High-risk AI systems require documented risk management, data quality protocols, technical documentation, automatic logging, human oversight mechanisms, and bias examination. Non-compliance fines run up to €35 million. A partner pitching to any EU-regulated organisation who does not raise this unprompted should be treated as insufficiently prepared.

 

🚩 Red Flag What it signals
Only PoC case studies — no production example covering monitoring, drift management, and retraining The partner operates as a data science consultancy, not an MLOps partner
“We can add MLOps later” The most expensive sentence in AI project management; retrofitting MLOps costs multiples of building it in from the start
Monitoring scope limited to infrastructure metrics only (CPU, memory, latency) Model quality degradation will go undetected until a business stakeholder complains
No answer to: “How do you detect and respond to data drift in production?” No operational MLOps practice exists — only theoretical knowledge
No EU AI Act awareness when pitching to organisations with EU operations The partner is not current on regulatory obligations that carry fines of up to €35 million

Criterion 4 — Delivery approach, governance, and handover

“Agile methodology” is table stakes — it tells you nothing useful. The specific pre-project artefacts a partner produces during discovery reveal far more about delivery discipline.

Non-negotiable before a project starts:

  • Architecture diagram of current state and proposed changes
  • High-level orchestration logic and data validation approach
  • Phased milestone plan: discovery → MVP → initial deployment → full rollout, with business KPIs at each gate
  • Security approach: PII handling, encryption standards, RBAC, compliance framework alignment
Delivery model Best fit Cost profile Management burden
Staff augmentation Specific skill gap, strong internal leadership Lower direct cost Client manages everything
Dedicated team Complex, long-term engagement; limited internal PM capacity Higher upfront, lower long-term TCO Vendor manages execution
Fixed scope Well-defined deliverables, fixed budget Predictable Risk: change-order disputes

 

The dedicated team model typically delivers lower long-term total cost of ownership: the vendor absorbs recruitment, management, and stabilisation costs, and reduced staff turnover translates directly into code consistency and lower onboarding overhead.

Many vendors treat go-live as the finish line. A launch that leaves the internal team unable to maintain or extend what was built is a delivery failure. Evaluate handover quality in advance:

  • “Can we speak to an internal engineering team you have handed a platform to in the last 12 months?”
  • “What does your post-deployment support model look like, and what is included versus extra cost?”
  • “How do you monitor cost anomalies and technical debt accumulation after go-live?”

Criterion 5 — Security, compliance, and data governance

Governance requirements are moving from good practice to legal obligation. Mature governance is what makes data infrastructure auditable, defensible, and extensible. In practical terms, this means a partner who:

  • Can demonstrate full data lineage — where data came from, what touched it, where it went
  • Has a defined approach to PII detection, masking, and minimisation
  • Uses RBAC as a default architectural choice, not an afterthought
  • Can show a real audit trail built for a regulated client — an artefact, not a marketing claim
  • Understands the difference between compliance (checkboxes) and governance (sustainable practice)

Ask for specific compliance evidence. A partner that has migrated GDPR-impacted data or built a HIPAA-compliant feature store should have concrete artefacts: architecture documentation, anonymisation approach, audit log schema.

 

🚩 Red Flag What it signals
“Our cloud provider takes care of security” Fundamental misunderstanding of the shared responsibility model — the partner is not accountable for application-layer security
No documented policies for PII handling, encryption at rest and in transit, RBAC Security is not operationalised; it exists only as a slide in the sales deck
Vague answers on incident response timelines and escalation paths No real incident response process exists; the client will be left managing breaches alone

Criterion 6 — Team composition and continuity

The leverage model common in enterprise systems integrators is one of the most important hidden risks in enterprise data partner selection. A partner sells the work, a senior manager oversees multiple concurrent projects, and a large base of junior associates does the bulk of execution.

The result is inconsistent quality and high turnover in the people actually doing the work.

What good team composition looks like

  • Data architect — designs the flow, structure, and storage strategy; must be directly accessible, not filtered through account management
  • Pipeline / data engineer(s) — senior-led, with platform-specific expertise in the client’s stack
  • Analytics engineer — dbt layer, semantic modelling, business logic ownership
  • MLOps engineer — if ML workloads are in scope; a distinct role from data engineer
  • QA lead — data validation, testing pipelines, quality gates before promotion
  • Project manager — milestone tracking, blocker management, stakeholder communication

Enterprise vs. boutique — when each makes sense:

Scenario Recommended partner type
Fortune 500, multi-year digital transformation, 50+ stakeholders, multi-country compliance Enterprise (Accenture, Deloitte, Capgemini)
Specific technical objective: migrate Teradata to Snowflake, build MLOps pipeline for claims fraud Boutique specialist — delivers 40–60% faster at comparable TCO
Mid-market company scaling toward enterprise Boutique with enterprise delivery methodology

 

The 40–60% delivery speed advantage for boutiques on defined technical objectives is noted in partner landscape research by DataEngineeringCompanies.com.

 

🚩 Red Flag What it signals
“Our cloud provider takes care of security” Fundamental misunderstanding of the shared responsibility model — the partner is not accountable for application-layer security
No documented policies for PII handling, encryption at rest and in transit, RBAC Security is not operationalised; it exists only as a slide in the sales deck
Vague answers on incident response timelines and escalation paths No real incident response process exists; the client will be left managing breaches alone

Criterion 7 — Commercial model, cost ownership, and exit terms

The financial structure of a partnership is a proxy for alignment. Partners who propose outcome-based pricing or shared milestones signal that their success is tied to the client’s. Pure time-and-materials billing with no performance thresholds creates less incentive to move efficiently.

Cloud cost design is a selection criterion, not an afterthought

Data engineering at scale generates substantial cloud spend. FinOps automation reduces cloud expenses by an average of 20%, and reserved instances and savings plans can deliver 25–70% compute cost savings depending on workload profile. A partner who does not discuss FinOps during architecture review — tagging strategy, right-sizing, reserved versus spot instance allocation — is either inexperienced at enterprise scale or indifferent to total cost.

On exit terms specifically

Many engagement contracts are well-defined on entry and entirely vague on exit. Before signing, establish in writing: who owns the code, pipelines, and models built during the engagement; what the offboarding process looks like; how data portability works; and what the notice period is. These are not adversarial questions — a partner who cannot answer them clearly has not thought through the end of the relationship, which is a signal about how they will manage the middle of it.

Hidden costs to build into the TCO model:

  • Strategic partnership initial investment: $100K–$500K depending on scope
  • Annual ongoing costs: 15–25% of initial investment
  • Data engineering typically represents 25–40% of total AI project spend
  • Model maintenance (retraining, drift detection, patching) adds 15–30% to operational costs annually
  • 85% of organisations misestimate AI project costs by more than 10% — contingency should be built in from day one
🚩 Red Flag What it signals
No visibility into rate cards or team cost breakdown Lack of transparency on cost structure makes TCO modelling impossible
Upsell-driven tooling decisions — recommending licences they resell Financial incentive misalignment; platform recommendations serve the partner, not the client
Deep discounts for long commitments before any pilot or proof of value Lock-in before trust has been established; the discount is the exit cost in disguise
Vague language around IP ownership of code, pipelines, and models The client may not own what was built with their own data and budget
No exit provisions — offboarding process and data portability undefined Vendor lock-in is structural; switching costs have been embedded into the engagement by design

3. The vendor shortlisting workflow

Five steps from longlist to signed contract, each with specific exit criteria.

Step 1 — Build the longlist (5–10 firms)

Start with cloud partner directories — AWS Partner Network, Google Cloud Partner Advantage, Snowflake Partner Connect, Databricks Partner Connect — then add analyst reports and trusted peer referrals. Filter out pure body-leasing shops if design authority is needed, and pure strategy firms if execution is what matters.

Step 2 — Run a structured discovery call

Send each vendor a one-pager before the call covering: current architecture (tools, sources, volumes), the three most critical business outcomes for the engagement, the internal team’s technical capacity, and any compliance or regulatory requirements. Use a consistent question set across all vendors, varying only for vendor-specific follow-ups. Inconsistent questioning makes comparison unreliable.

Step 3 — Score on a 1–5 scale per criterion

Complete scoring before any commercial conversation. Apply weights to reflect the situation.

Data Engineering - Business Evaluation

Step 4 — Run a bounded pilot

A well-designed pilot is the single best risk-management tool available. Design parameters:

  • Duration: 4–8 weeks maximum, with weekly milestone checkpoints — not just a final demo
  • Scope: one well-defined pipeline or data product tied to a live business metric
  • Constraint: use existing data sources and tools — avoid introducing new platforms during validation
  • Budget: fixed and agreed upfront, with explicit criteria for stopping if milestones are missed

What to judge beyond technical delivery

The quality of discovery questions asked during design; documentation produced (architecture decisions log, data dictionary, deployment runbook); communication behaviour under unexpected complexity; how the delivery team interacted with internal engineers; and whether the internal team could maintain and extend what was built.

Step 5 — Decide on partnership structure

Once the pilot is complete and a vendor is selected, define the ongoing engagement model. A dedicated squad suits complex, long-horizon projects. Flexible capacity suits organisations with strong internal teams that need specialist uplift. Advisory-only works where the internal team can execute but needs strategic guidance.

The signal that a vendor relationship has matured into a genuine strategic partnership: joint ownership of the data roadmap, shared governance forums, and aligned KPIs. Until those elements are in place, the relationship is that of a service provider.

4. How this plays out in practice

The head of data at a mid-market European insurer sent the same RFP to nine vendors. Eight came back with near-identical decks: a Snowflake logo, a GDPR slide, a timeline. The ninth had found her company’s SQL Server schema in a public regulatory filing and built their response around it. That vendor made the shortlist before the first call.

The company needed two things done simultaneously:

  • Migrate a claims warehouse from SQL Server to Snowflake
  • Take ownership of three ML models running in production with no monitoring, no documentation, and an EU AI Act deadline closing in

Three vendors made the final round.

The large European had the strongest credentials on paper. They also proposed a twelve-week discovery phase before writing a line of code. When pressed on the EU AI Act timeline, every specific question was escalated to a specialist who replied four days later — with a response that would have fitted any financial services client.

The boutique data engineering firm was faster and sharper technically. Their reference client gave a strong recommendation. The problem surfaced in the third conversation, when the head of data asked how they would approach the three existing ML models. The lead engineer said they could document what was there and hand it to the data science team.

There was no data science team. That answer ended their candidacy — not because it was wrong, but because it showed they had not absorbed the problem.

The third vendor was smaller and had never done a migration at this scale. What they brought instead:

  • A production fraud model running at a Belgian insurer
  • Drift monitoring, a documented EU AI Act risk assessment, and a retraining trigger tied to PSI thresholds
  • The Belgian insurer’s engineering lead on the discovery call — unprompted

It was the only conversation in the process where something genuinely useful was said without being asked for.

The pilot ran six weeks. On week three, the third vendor flagged that one of the existing models was pulling a feature — policyholder postcode — from a table remapped two months earlier in the legacy system. The model was live. Its predictions were wrong. Nobody had noticed.

That finding won the engagement. Not the credentials, not the pricing, not the migration track record. The vendor who found a problem the client did not know they had.

DataEngineering Selection Framework

Summary

BCG’s 2024 research across more than 1,000 companies worldwide found that organisations that had moved beyond proof of concept and were generating tangible AI value achieved 1.5x higher revenue growth, 1.6x greater shareholder returns, and 1.4x higher return on invested capital over the prior three years compared to peers. The research identified a consistent differentiator: those companies prioritised data quality and data management as foundational infrastructure, and they built governance and operational capability before scaling AI. The technology itself was not what separated them.

That infrastructure does not build itself, and it does not survive a poor partner selection. The framework in this guide exists to close the distance between the decision being made today and the outcome that depends on it.


FAQ


What’s the difference between a data engineering company and an MLOps partner?

plus-icon minus-icon

A data engineering partner builds the foundation — pipelines, warehouses, data flow. An MLOps partner takes it further into machine learning: training models, deploying them, monitoring performance, and keeping them up to date.
Many vendors say they do both — but the real test is whether they’ve actually run models in production, with monitoring and retraining in place.


How do I know where we are right now?

plus-icon minus-icon

  • Greenfield: everything is manual, no real infrastructure yet
  • Scaling: you have systems, but they’re breaking, slowing down, or getting expensive
  • AI-curious: models work in notebooks, but not reliably in production

Most companies in the third group are also dealing with issues from the second.


Big consultancy or boutique specialist?

plus-icon minus-icon

If you’re running a massive, multi-country transformation — go with a large consultancy.
If you have a clear technical goal (like a migration or ML pipeline), boutiques are usually faster and just as cost-effective.
Mid-sized companies? Typically better off with a strong boutique partner.


What about EU AI Act compliance?

plus-icon minus-icon

If you operate in the EU and use AI for decisions (like fraud, hiring, credit scoring), you’re likely in “high-risk” territory. Rules kick in from August 2026.

Your partner should understand things like risk management, data quality, logging, bias checks, and human oversight.
If they don’t — that burden lands on you.


How do we assess governance maturity?

plus-icon minus-icon

Ask for proof, not promises.
They should be able to show things like:

  • data lineage examples
  • anonymisation approaches
  • audit logs
  • access control policies

If they can’t — that’s a red flag.


What’s the biggest hidden cost?

plus-icon minus-icon

Ongoing model maintenance — retraining, monitoring, fixing — can add 15–30% annually and is often overlooked.

Second: cloud costs.
Without good FinOps practices, you’ll end up with a system that works… but is way too expensive to run.


When does a vendor become a strategic partner?

plus-icon minus-icon

When three things happen:

  • you share a data roadmap
  • you have joint governance
  • your KPIs are aligned

Until then, it’s just a service relationship — which is fine, as long as you treat it that way.




Category:


Data Engineering