
There is a particular kind of meeting that happens too often in organizations adopting AI. The vendor demo was impressive. The pilot showed promise. The contract was signed. And now, six months later, the team is discovering that their data is being used to train the vendor's models, that pricing has ballooned beyond projections, that switching to an alternative would require rebuilding from scratch, and that the "enterprise-ready" solution requires far more customization than anyone anticipated.
This is not a failure of technology. It is a failure of evaluation.
AI vendor selection differs fundamentally from traditional software procurement. The risks are different, the questions that matter are different, and the contract terms that protect you are different. Organizations that apply conventional vendor evaluation frameworks to AI purchases consistently underestimate these differences, and pay the price.
AI procurement is risk management, not feature comparison. The questions that protect you are the ones vendors hope you won't ask.
I want to share what we have learned about evaluating AI vendors effectively. Not the sanitized checklist that could apply to any technology purchase, but the specific questions, red flags, and protective measures that matter for AI.
In This Post
- The Questions Vendors Hope You Won't Ask
- Red Flags That Should Stop You Cold
- The Contract Clauses That Actually Matter
- The Evaluation Process That Works
- What Good Looks Like
- The Leverage You Have
The Questions Vendors Hope You Won't Ask
The most important questions in AI vendor evaluation are often the ones that sales teams are trained to deflect. Ask them anyway.
"What happens to my data?"
This is the question that separates sophisticated buyers from those who will regret their decisions. Many AI vendors claim broad data usage rights in their terms of service, often exceeding what traditional SaaS providers request.[1] Many vendors default to using customer data to improve their models unless the contract explicitly prohibits it.
You need specific answers:
- Will your data be used to train the vendor's general-purpose models?
- What about embeddings, logs, and fine-tuning outputs?
- How long is data retained after the engagement ends?
- Can you get your data deleted on request?
Vague assurances like "we take privacy seriously" are not answers.
"How was your model trained?"
Understanding training data matters for compliance, ethics, and legal exposure. Ask about data sources, verification processes, and bias mitigation. If the vendor cannot or will not explain their training data provenance, that opacity may eventually become your problem through regulatory scrutiny or reputational risk.[2]
"What happens when your model changes?"
AI vendors frequently update their models, sometimes silently. These updates can alter behavior in ways that affect your operations. Ask whether you will receive notice before model updates, whether you can test changes before they affect production, and whether you can pin to specific model versions.
"Same endpoint, new behavior" is an operational risk that traditional software evaluation does not address.
"How do I leave?"
Vendor lock-in in AI can be severe. Your data, your fine-tuning work, your integrations may all become difficult to migrate. Ask about data export in open formats, assistance with migration, and what happens to your data after contract termination. If the exit path is unclear or expensive, factor that into your total cost of ownership.
"Who else touches my data?"
Many AI vendors are front-ends to larger APIs and cloud providers. Your data may flow through multiple sub-processors you have never heard of. Ask for a complete list of sub-processors, advance notice before new ones are added, and flow-down of security and privacy terms to all parties handling your data.
Red Flags That Should Stop You Cold
Certain warning signs indicate deeper problems that will likely surface after contract signing. Treat these as serious concerns, not minor irritants.
| Red Flag | What It Signals | What to Do |
|---|---|---|
| Missing or boilerplate documentation | Not prepared for enterprise deployment | Request AI-specific policies before proceeding |
| Cannot produce compliance evidence | Certifications may be lapsed or exaggerated | Require current SOC 2 report within two weeks |
| Evasive security responses | Hiding problems or lacking maturity | Insist on concrete architecture details |
| No industry references | May not understand your domain's requirements | Ask for references you can actually call |
| Pricing opacity | Budget overruns likely | Demand clear projections under realistic usage |
| Won't discuss limitations | Hiding problems or doesn't understand own tech | Consider this disqualifying |
Missing or boilerplate documentation
A trustworthy AI vendor will have comprehensive, specific policies: privacy policies, data processing agreements, security documentation, incident response plans. If these documents are missing, vague, or clearly copied from a template without customization for AI-specific concerns, the vendor is not prepared for enterprise deployment.
Inability to produce compliance evidence
If a vendor claims SOC 2 compliance but cannot produce a current audit report within a week or two, something is wrong. SOC 2 Type II reports typically cover a 12-month period; reports significantly older than that may indicate lapsed certification.[3] If the vendor deflects requests for documentation until after procurement, expect problems.
Evasive responses to security questions
"We take security seriously" is not an answer. "Our proprietary security prevents disclosure" is a red flag. Legitimate enterprise vendors can explain their security architecture, encryption standards, access controls, and incident response procedures in concrete terms.
Pricing opacity
AI pricing models are notoriously complex. Usage-based pricing can spiral unpredictably. Token costs for RAG implementations can balloon as context sizes grow.[4] If the vendor cannot provide clear cost projections under realistic usage scenarios, budget overruns are likely.
The Contract Clauses That Actually Matter
Standard software contracts are insufficient for AI engagements. Several provisions require specific attention.
Data rights and training restrictions
This is non-negotiable. Your contract should explicitly state that the vendor cannot use your data, prompts, outputs, or derived data to train models that serve other customers. This restriction should survive contract termination. Without explicit prohibition, assume your data will be used.[1]
Data ownership and portability
Confirm in writing that you own your input data, output data, and any fine-tuning or customization work. Require that the vendor provide data export in standard, open formats upon request and assist with migration at reasonable cost. Specify that all copies of your data must be deleted after the relationship ends.
Model versioning and change notification
Require advance notice before model updates that could affect your operations. Negotiate the right to test changes before they reach production. Consider whether you need the ability to pin to specific model versions for stability.
Performance standards and degradation
AI systems produce probabilistic outputs that can degrade over time. Your contract should define minimum accuracy thresholds, specify vendor obligations to address performance degradation, and establish how performance will be measured. Include service level agreements with meaningful remedies for failures.
Liability and indemnification
Traditional software liability caps may be insufficient for AI-specific risks: hallucinated outputs used in business decisions, intellectual property claims stemming from training data, regulatory violations from automated decisions.[5] Push for vendor indemnification covering data breaches, IP infringement, and regulatory violations. Consider whether consequential damages from AI errors should be recoverable.
Sub-processor transparency
Require a complete list of all parties who will handle your data, advance notification before new sub-processors are added, and the right to object or terminate if sub-processor changes create unacceptable risk. Ensure that all your contractual protections flow down to sub-processors.
Exit provisions
Specify what happens at contract termination: data export timeline, deletion confirmation, transition assistance, and any fees associated with departure. The time to negotiate exit terms is before you sign, not when you are trying to leave.
The Evaluation Process That Works
Structure matters. A haphazard evaluation process produces haphazard results.
-
Start with requirements, not vendors. Before evaluating any vendor, document what you actually need: the business problem you are solving, the data you will be processing, the accuracy and latency requirements, the regulatory constraints, the integration points with existing systems. Vendors are skilled at redirecting conversations toward their strengths. Clear requirements keep the focus on your needs.
-
Build a cross-functional evaluation team. AI vendor evaluation requires perspectives that technical teams alone cannot provide. Include procurement for contract terms, legal for liability and compliance, security for data protection, and business stakeholders for use case fit. The questions that matter most often come from outside the technical domain.
-
Request evidence, not claims. Vendors will make claims about accuracy, security, and enterprise readiness. Your job is to verify those claims. Request case studies with measurable outcomes in your industry. Ask for customer references you can actually contact. Require compliance documentation before purchase, not after. If a vendor resists providing evidence, treat that resistance as information.
-
Conduct realistic pilots. Pilot projects should use representative data, realistic volumes, and production-like conditions. A demo that works on curated examples may fail with your actual data. Define success criteria before the pilot begins, and measure rigorously against those criteria.
-
Calculate total cost of ownership. The license fee or API cost is the beginning, not the end, of your cost analysis. Include integration expenses, data preparation, training and change management, ongoing maintenance, and potential cost growth as usage scales. Build scenarios for both expected and elevated usage levels.
-
Document everything. Keep records of vendor claims, pilot results, reference conversations, and negotiation discussions. If disputes arise later, documentation protects you. If the vendor's behavior changes after contract signing, documentation establishes what was promised.
What Good Looks Like
To be fair, strong AI vendors do exist, and they share common characteristics.
-
Transparency about capabilities and limitations. Good vendors explain what their technology does well and where it struggles. They provide documentation about training data, model architecture, and known failure modes. They do not oversell.
-
Clear, specific policies. Data handling, security practices, compliance certifications: these should be documented, current, and available before purchase. The documentation should address AI-specific concerns, not just general software security.
-
Willingness to negotiate. Even vendors with standard terms will typically negotiate for enterprise deals, particularly on data rights and liability. A vendor that refuses any negotiation may be signaling how they will behave when problems arise.
-
Proactive communication. Good vendors provide notice before changes, respond promptly to questions, and surface potential issues rather than hiding them. The relationship during evaluation is a preview of the relationship during implementation.
-
References who speak candidly. When you contact references, they should be able to discuss challenges as well as successes. If every reference sounds like a marketing testimonial, you are not getting the full picture.
The Leverage You Have
Buyers often underestimate their negotiating position. AI vendors need customers, particularly enterprise customers who provide revenue stability, case study potential, and market credibility. You have more leverage than you think.
Use that leverage to protect yourself. Push back on one-sided terms. Request modifications to standard contracts. Walk away from vendors who will not engage seriously with your concerns. The vendors who value your business will work with you. The vendors who will not are revealing something important about how they operate.
The AI vendor landscape is crowded and competitive. If one vendor will not meet your requirements, others likely will. The scarcity is on the vendor side, not the buyer side.
A Closing Thought
Vendor evaluation is fundamentally about risk management. You are not trying to find a perfect vendor; you are trying to identify a vendor whose capabilities match your needs and whose risks you can manage.
There are some things we can do to manage those risks effectively:
- Ask the questions that matter, even when vendors would prefer we did not
- Demand evidence for claims rather than accepting assertions
- Structure contracts that protect our interests and provide recourse when things go wrong
- Build evaluation processes that surface problems before contracts are signed rather than after
The goal is not to approach vendors with suspicion. Many AI vendors are building valuable technology and operating with integrity. The goal is to approach vendor selection with the rigor that the decision deserves. AI systems will touch your data, influence your operations, and affect your customers. That warrants careful evaluation.
The organizations that get AI vendor selection right are not those who find vendors without flaws. They are those who understand the flaws before signing, negotiate appropriate protections, and enter relationships with clear expectations on both sides.
That clarity starts with asking better questions.
This is the ninth in our January series on data and AI strategy for 2026. Subscribe to receive the full series as it publishes throughout the month.
Sources
-
IAPP (International Association of Privacy Professionals). AI Governance in Practice Report (2024) on enterprise governance and vendor data practices. iapp.org
-
EU AI Act, Article 10 on training data transparency requirements. artificialintelligenceact.eu
-
AICPA. SOC 2 Reporting guidance on audit periods and report validity. aicpa-cima.com
-
OpenAI pricing documentation. Token-based pricing models and context window costs. openai.com
-
Gartner. Research on AI trust, risk, and security considerations in enterprise procurement. gartner.com