Ready to turn insight into action?
We help organisations transform ideas into measurable results with strategies that work in the real world. Let’s talk about how we can solve your most complex supply chain challenges.
How to Evaluate Tenders: Methodology, Common Mistakes, and What Good Looks Like in Australia
The tender evaluation is the point in the procurement process where the investment in planning, specification, and market engagement either pays off or falls apart. A well-run evaluation takes well-structured submissions from capable suppliers and applies a rigorous, transparent methodology to identify the one that offers the best value for money. A poorly run evaluation takes the same submissions and produces a decision that is either wrong (selecting a supplier that does not represent the best outcome) or indefensible (selecting the right supplier through a process that cannot withstand challenge).
Both outcomes are common in Australian procurement. The wrong supplier gets selected because the evaluation criteria were poorly designed, the scoring was inconsistent, or the panel did not have the expertise to assess the submissions properly. The right supplier gets selected but through a process so poorly documented or so procedurally flawed that an unsuccessful tenderer could challenge the outcome, and in the public sector increasingly does challenge it.
This article covers how to design and execute a tender evaluation that produces a good decision and a defensible process. It is written for procurement practitioners, evaluation panel members, and the managers who approve procurement recommendations in both public and private sector organisations.
Designing the Evaluation Before Writing the Tender
The evaluation methodology should be designed before the tender documentation is written, not after submissions are received. This is a fundamental principle that is violated more often than it is observed.
The reason is straightforward. The evaluation criteria, weightings, and scoring methodology determine what information you need from tenderers. If you design the evaluation first, you can structure the tender documentation to elicit exactly the information you need to evaluate against your criteria. If you write the tender first and design the evaluation later, you will almost certainly find that the submissions do not contain the information you need, or contain it in a format that is difficult to assess consistently.
In Australian government procurement, this principle is not optional. Under the Commonwealth Procurement Rules (effective November 2025), evaluation criteria must be stated in the request documentation. State government procurement frameworks have equivalent requirements. The evaluation methodology must be determined and documented before the approach to market is released.
In private sector procurement, there is no legislative requirement, but the discipline of designing the evaluation before the tender applies equally. Organisations that skip this step consistently produce weaker evaluation outcomes.
Evaluation Criteria Design
The evaluation criteria are the dimensions against which submissions will be assessed. Getting the criteria right is the single most important design decision in the evaluation process.
Relevance. Every criterion must be relevant to the procurement outcome. A criterion that does not help distinguish between submissions or does not relate to the value the organisation is seeking from the procurement should not be included. Evaluation panels sometimes include criteria because they seem important in the abstract (innovation, sustainability, corporate social responsibility) without considering whether they are genuinely differentiating for this specific procurement. Every criterion adds assessment burden. If it does not add assessment value, remove it.
Completeness. The criteria, taken together, should cover all the dimensions that matter for the procurement decision. If price, technical capability, experience, transition approach, and risk management all matter, they should all be represented. Missing a dimension that turns out to be important post-award (the supplier had no relevant experience, the transition plan was unrealistic) is a failure of evaluation design.
Assessability. Each criterion must be assessable based on the information that tenderers will provide. If you include "demonstrated experience in complex logistics operations" as a criterion, the tender documentation must ask tenderers to provide information that allows the panel to assess this: case studies, reference sites, project descriptions. A criterion that cannot be assessed from the submission is a criterion that will be scored on impression rather than evidence.
Independence. The criteria should be as independent of each other as possible. If "technical capability" and "relevant experience" are so closely related that the panel would score them based on the same information, they should be combined into a single criterion or clearly differentiated in the assessment guidance. Overlapping criteria create double-counting, where a strong (or weak) aspect of a submission is rewarded (or penalised) twice.
Number. Less is more. Five to eight evaluation criteria is typically the right range for a complex procurement. Fewer than five risks missing an important dimension. More than eight creates assessment fatigue and dilutes the weight of each criterion to the point where none of them is truly differentiating.
Weighting
Each criterion should be assigned a percentage weighting that reflects its relative importance to the procurement outcome. The weightings must add to 100% and must be disclosed to tenderers in the request documentation.
The weighting decision is a strategic choice that signals to the market what the organisation values. A procurement weighted 60% on price and 40% on non-price criteria tells the market that cost is the dominant consideration. A procurement weighted 40% on price and 60% on non-price criteria tells the market that capability, experience, and approach matter more than being the cheapest.
There is no universally correct weighting. It depends on the procurement. A commodity purchase where specifications are fixed and quality is assured might reasonably weight price at 70% or higher. A complex services engagement where the quality of the team, the methodology, and the relationship will determine the outcome might weight price at 30% or lower.
The common mistake is defaulting to a standard weighting (typically 60/40 non-price/price) without considering whether that weighting is appropriate for the specific procurement. The weighting should be a deliberate decision, made during evaluation design, that reflects what genuinely matters.
Price weighting deserves particular attention. Weighting price too heavily in a services procurement incentivises tenderers to submit the lowest price they can justify, which may mean under-resourcing the engagement, deploying junior staff, or cutting scope. The organisation gets a low price and a poor outcome. Weighting price too lightly removes the competitive tension that drives value for money. The right balance ensures that price is a genuine differentiator without being the only differentiator.
Scoring Methodology
The scoring methodology defines how each criterion is assessed and scored. Two approaches are common in Australian procurement.
Qualitative scoring against a defined scale. Each criterion is assessed qualitatively, with the panel assigning a score based on a defined scale. A common scale uses five or six levels, from "does not meet requirements" to "significantly exceeds requirements," with each level defined by descriptors that guide the panel on what constitutes performance at that level. The panel assesses each submission against each criterion, discusses the assessment, and agrees a consensus score. This approach works well for criteria that are inherently qualitative (methodology, experience, team capability, approach) and is the most common approach in Australian procurement.
The quality of this approach depends entirely on the quality of the scoring descriptors. Descriptors that are vague ("good response," "strong capability") produce inconsistent scoring because different panel members interpret them differently. Descriptors that are specific ("demonstrated experience in at least three comparable projects within the last five years, with verified outcomes") produce more consistent and defensible scoring.
Quantitative scoring with a formula. For criteria that can be measured objectively, particularly price, a formula-based approach converts the raw data into a score. The most common price scoring formula in Australia normalises tenderer prices against the lowest price, so the lowest price receives the maximum score and higher prices receive proportionally lower scores. Several formula variations exist, and the choice of formula can materially affect the outcome, particularly when the price range across tenderers is wide. The formula should be selected during evaluation design and disclosed to tenderers.
For most procurements, a hybrid approach works best: qualitative scoring for non-price criteria and quantitative scoring for price, with the results combined using the pre-determined weightings to produce a total weighted score.
The Evaluation Panel
The composition of the evaluation panel determines the quality of the evaluation. The panel should include people who collectively have the expertise to assess all of the evaluation criteria, the authority to make or recommend a procurement decision, and the objectivity to assess submissions fairly.
Technical expertise. At least one panel member should have deep expertise in the subject matter of the procurement. For a logistics tender, this means someone who understands logistics operations. For an IT systems tender, someone who understands the technology. Without technical expertise on the panel, the evaluation of capability and methodology criteria will be superficial.
Procurement expertise. At least one panel member should understand procurement process, evaluation methodology, and probity requirements. This person ensures that the evaluation is conducted in accordance with the methodology, that scoring is consistent and evidence-based, and that the process is properly documented.
End-user perspective. Where the procurement is for a service or product that will be used by a specific business unit, a representative of that business unit should be on the panel. They bring the perspective of the people who will live with the outcome of the procurement decision.
Independence. Panel members must not have conflicts of interest with any of the tenderers. In government procurement, this is a formal requirement with declaration and management protocols. In private sector procurement, it is equally important for the integrity of the process.
Panel size. Three to five panel members is the typical range. Fewer than three risks insufficient perspective and makes consensus scoring vulnerable to individual bias. More than five makes scheduling difficult, extends the evaluation timeline, and can dilute accountability.
Running the Evaluation
The evaluation itself should follow a structured process.
Individual assessment. Each panel member reads and assesses each submission independently, before the panel meets to discuss. This ensures that every panel member forms their own view before being influenced by others. Individual assessment sheets, recording each member's scores and notes against each criterion, provide the evidence base for the subsequent consensus discussion.
Consensus scoring. The panel meets to discuss each criterion for each submission, compare their individual assessments, and agree a consensus score. Where panel members have scored differently, the discussion should focus on what evidence in the submission supports each assessment. The consensus score should reflect the panel's collective view, not an average of individual scores. The discussion and the rationale for the consensus score should be documented.
Clarifications. If the panel identifies gaps, ambiguities, or inconsistencies in a submission that affect the assessment, clarification should be sought from the tenderer. Clarifications must be managed carefully to maintain fairness: the same opportunity must be available to all tenderers, clarification questions should not coach or lead the tenderer toward a better response, and the clarification process should be documented.
Price assessment. Price should be assessed separately from non-price criteria, ideally by a different subset of the panel or at a different point in the process, to prevent price information from influencing the assessment of non-price criteria. Price is assessed for completeness (does it cover the full scope?), competitiveness (how does it compare to other submissions and to the pre-tender estimate?), and sustainability (is the price realistic, or is it likely to result in variations, scope reduction, or under-resourcing?).
Moderation and calibration. Before finalising scores, the panel should review the overall results for internal consistency. Does the ranking make sense? Are the scores for each criterion consistent across tenderers (is the scoring scale being applied consistently)? Are there any anomalies that suggest a scoring error or a misunderstanding of the criteria? This moderation step catches errors and improves the quality of the final assessment.
Common Mistakes
Scoring on impression rather than evidence. The evaluation should be based on what is in the submission, not on what the panel believes about the tenderer from prior experience or reputation. If a tenderer is well known and highly regarded but has submitted a poor response, the response should score poorly. If an unknown tenderer has submitted an exceptional response, it should score highly. The evaluation assesses the submission, not the tenderer's reputation.
Anchoring on price. When the panel knows the prices before assessing non-price criteria, there is a documented cognitive bias toward adjusting non-price scores to justify selecting the lowest price. This is why price should be assessed separately, and ideally after non-price scoring is complete.
Inconsistent scoring across tenderers. The scoring scale needs to be applied consistently. If one tenderer receives a score of 8 out of 10 for "demonstrated three relevant projects," another tenderer who has also demonstrated three comparable projects should receive a similar score. Consistency does not mean identical scores, as the quality of the projects and the evidence provided may differ, but the same standard should be applied to all submissions.
Inadequate documentation. The evaluation record, including individual assessment sheets, consensus scoring rationale, clarification correspondence, and the evaluation report, is the evidence base for the procurement decision. If the decision is challenged, internally or by an unsuccessful tenderer, the documentation must demonstrate that the evaluation was conducted in accordance with the stated methodology, that scores were based on evidence, and that the process was fair and transparent. Inadequate documentation turns a good evaluation into a vulnerable one.
Changing the methodology mid-evaluation. Once the evaluation criteria, weightings, and scoring methodology have been set and disclosed to tenderers, they should not be changed. If the panel discovers mid-evaluation that a criterion is not useful or that the weighting is producing anomalous results, the correct response is to document the issue and manage it through the moderation process, not to change the rules after submissions have been received.
Over-reliance on presentations. Some evaluation processes include a presentation or interview stage where shortlisted tenderers present to the panel. This can be valuable for assessing team capability and cultural fit. But the presentation should be scored against defined criteria, not used as an opportunity for the tenderer to override a weak written submission with a polished pitch. The written submission should carry the majority of the weight, because it represents the tenderer's considered, documented response.
Government-Specific Considerations
Australian government procurement has specific evaluation requirements that practitioners need to understand.
Under the Commonwealth Procurement Rules, evaluation criteria must be stated in the request documentation, value for money must be the primary consideration, and non-compliance with the evaluation process can create legal exposure under the Government Procurement (Judicial Review) Act 2018. The revised CPRs effective November 2025 also require that non-price factors, including ethical conduct, labour compliance, and environmental impact, be considered in value-for-money assessments.
State government frameworks have equivalent requirements, though the specific rules vary by jurisdiction. Victorian Government Purchasing Board policies, NSW Procurement Policy Framework, Queensland Procurement Policy, and Western Australian State Supply Commission policies each set out evaluation requirements that must be followed by agencies in those jurisdictions.
For government procurements, the evaluation is not just a decision-making tool. It is a compliance obligation. The process, the documentation, and the outcome must all be defensible against audit scrutiny and potential legal challenge.
How Trace Consultants Can Help
Trace supports Australian organisations across the full procurement evaluation lifecycle.
Evaluation framework design. We design evaluation criteria, weightings, and scoring methodologies that are tailored to the specific procurement, defensible under the relevant regulatory framework, and structured to produce a clear differentiation between submissions.
Evaluation panel support. We provide specialist panel members for complex procurements, bringing deep category expertise, procurement process knowledge, and experience across hundreds of evaluations in both public and private sectors.
Evaluation facilitation. We facilitate consensus scoring sessions, ensuring that the process is rigorous, consistent, and properly documented. Our facilitation keeps panels on track, on methodology, and focused on evidence.
Evaluation report and recommendation. We prepare evaluation reports that clearly document the process, the assessment, and the recommendation, providing the procurement decision-maker with the information and confidence needed to approve the outcome.
Explore our Procurement services →Explore our Government & Defence sector expertise →Speak to an expert at Trace →
Getting Started
If you are about to run a tender evaluation, the single most important thing you can do is design the evaluation before you write the tender. Define your criteria, set your weightings, write your scoring descriptors, decide your methodology, and brief your panel before a single submission is received. That upfront investment in evaluation design will pay for itself many times over in the quality of the decision and the defensibility of the process.
A procurement process is only as good as the evaluation that concludes it. The best specification, the most competitive field of tenderers, and the most thorough market engagement are all wasted if the evaluation does not produce the right decision through a robust process. The evaluation is where the value is realised. It deserves the same rigour as every other stage of the procurement.
Ready to turn insight into action?
We help organisations transform ideas into measurable results with strategies that work in the real world. Let’s talk about how we can solve your most complex supply chain challenges.







