Common mistakes
Technology vendor selection is one of the highest-impact decisions organisations make, and one of the most commonly botched. The wrong choice creates years of pain: systems that don't fit, vendors that can't deliver, costs that escalate well beyond the original proposal.
The mistakes that haunt organisations:
- Demo-driven decisions. Choosing based on the best sales presentation rather than actual fit. Every vendor's demo looks amazing. It's literally what they rehearse.
- Price over value. The lowest initial quote often means the highest total cost once implementation, customisation, and ongoing support are factored in.
- Ignoring implementation. The product is maybe 30% of success. How it gets implemented, the approach, the team, the methodology, determines the other 70%.
- Insufficient requirements. Vague requirements get vague proposals. Then you're surprised when the delivered system doesn't match expectations.
- Not involving users. IT selects a system; users reject it. This happens more often than anyone admits.
- Single-vendor shortlist. Confirming a predetermined choice isn't selection. It's ceremony.
Phase 1: Requirements definition
Before looking at any vendor, understand what you actually need. Clear requirements prevent scope creep during evaluation and ensure you're comparing like with like.
Business requirements
Start here, not with features. What business problems does this solve? What processes will change? Who are the actual users and what do they need from their perspective? What outcomes define success?
Functional requirements
What must the system do? What workflows does it need to support? What data does it handle? What integrations with existing systems are required? Be specific. "Must integrate with Xero" is useful. "Must integrate with our accounting" is not.
Non-functional requirements
- Performance: expected response times, throughput, concurrent users
- Scalability: growth expectations over the next 3–5 years
- Security: compliance requirements, data protection, access control
- Availability: uptime requirements, disaster recovery expectations
- Support: SLA requirements, support hours, Australian-timezone coverage
MoSCoW prioritisation
Not all requirements are equal. Categorise each as Must have, Should have, Could have, or Won't have (this time). Focus vendor evaluation on Must-haves. Use Should-haves as differentiators between otherwise similar vendors.
Phase 2: Shortlisting
Market research
- Analyst reports (Gartner Magic Quadrant, Forrester Wave) for market overview. Useful for orientation, but don't treat them as gospel
- Peer recommendations from industry contacts. Often the most reliable signal
- Online reviews (G2, Capterra). Useful for patterns, apply healthy scepticism to individual reviews
- Industry forums and communities
From long list to short list
Start with 6–10 potential vendors, then shortlist to 3–4 for detailed evaluation based on:
- Apparent capability to meet your Must-have requirements
- Vendor financial stability and viability
- Relevant industry experience and references
- Geographic coverage and support model (Australian support is valuable)
- Approximate fit with your budget range
Phase 3: RFP process
A Request for Proposal gives you structured, comparable information from each shortlisted vendor.
RFP components
- Company overview: your organisation, the project context, and what you're trying to achieve
- Requirements: the detailed functional and non-functional requirements
- Response format: how you want answers structured (makes comparison much easier)
- Commercial requirements: pricing format, contract terms, payment expectations
- Timeline: submission deadline, evaluation schedule, expected decision date
RFP tips: Be specific. Vague requirements get useless responses. Ask for case studies and references in your industry. Request detailed pricing that includes implementation costs. Include realistic scenarios for vendors to address. And always allow a questions period where vendors can clarify requirements before responding.
Phase 4: Evaluation
Weighted scoring
Weighted scoring ensures consistent, comparable evaluation across vendors:
| Category | Weight | What you're assessing |
|---|---|---|
| Functional fit | 30% | Coverage of Must-have requirements |
| Technical fit | 20% | Architecture, integration capability, security |
| Vendor viability | 15% | Financial stability, product roadmap, references |
| Implementation approach | 15% | Methodology, team quality, timeline realism |
| Total cost | 20% | TCO over 5 years, not just the initial price |
Demonstrations
Require scripted demonstrations based on your specific scenarios, not the vendor's canned demo. Have actual end users attend and provide feedback. Score against predefined criteria, not presentation polish.
Reference checks
Talk to actual customers. Ideally, find references the vendor didn't nominate. LinkedIn connections in similar industries are gold. Ask about the implementation experience, ongoing support quality, and what they'd do differently.
Proof of concept
For high-stakes selections, a paid proof of concept with 1–2 finalists validates real-world fit. Configure the system for your actual use cases with your actual data. The cost is small compared to choosing wrong and living with it for years.
Phase 5: Selection and negotiation
Final decision
Base it on the evaluation scores, reference feedback, and PoC results. Make sure key stakeholders are aligned before announcing. Disagreements that surface after the decision derail implementations.
Contract negotiation
- Lock in pricing terms for multi-year commitments
- Define SLAs with meaningful consequences for non-performance
- Tie implementation payments to milestones and acceptance criteria
- Ensure clear data ownership and data export rights
- Cap annual price increases (vendors love uncapped "CPI adjustments")
- Include termination rights and transition support obligations
Frequently asked questions
How long should the selection process take?
For significant technology decisions: 8–12 weeks from requirements to selection. Rushing leads to bad decisions; dragging it out loses vendor engagement and internal momentum. For smaller purchases, you can compress this significantly.
Should we use a consultant for vendor selection?
Independent advisory can add value when you lack domain expertise in the technology area, when the decision is high-stakes, or when internal politics might bias the outcome. Just make sure the consultant doesn't have commercial relationships with the vendors being evaluated.
What if no vendor meets all our requirements?
That's normal. No product is a perfect fit. Evaluate the gaps: are they in Must-have requirements (problem) or Could-have requirements (acceptable)? Can gaps be closed through configuration, customisation, or complementary tools? The best-fit vendor with a clear plan for addressing gaps usually beats waiting for perfection.
Key takeaways
- The wrong vendor choice creates years of pain: systems that don't fit, costs that spiral, implementations that fail.
- Start with clear requirements before looking at products. Vague requirements get vague proposals.
- The product is 30% of success; implementation is 70%. Evaluate the implementation approach, not just the software.
- Evaluate total cost of ownership over 5 years, not initial price. The cheapest option upfront is often the most expensive overall.