How Aptitude Tests Are Scored: Percentiles, Norm Groups, Cut-Offs and What Employers See

You have just finished an aptitude test and the screen shows a confirmation message. But what happens next? How does a test provider turn your answers into a score, and what does your employer actually see in the results? Understanding how aptitude tests are scored removes the mystery from the process and helps you interpret feedback, set realistic targets, and focus your preparation on what genuinely moves the needle.

Aptitude test scoring is more complex than simply counting correct answers. Employers like Google, Deloitte, Unilever, JP Morgan, and the Civil Service use sophisticated scoring methods that compare your performance against thousands of other candidates, adjust for test difficulty, and translate raw data into standardized metrics that support fair hiring decisions. This guide explains every layer of that process so you know exactly where you stand.

How Norm-Referenced Scoring Works

The foundation of almost all aptitude test scoring is norm-referenced scoring. Rather than grading you against an absolute standard, the test provider compares your performance to a reference population known as the norm group. Your final score reflects where you fall within that distribution.

Here is how the process works step by step. First, the test provider administers the assessment to a large, representative sample of people. This sample becomes the norm group. The provider then calculates the statistical distribution of scores across that group, identifying the mean, standard deviation, and percentile boundaries. When you take the test, your raw score is mapped onto that distribution. The result is a percentile rank that tells the employer how your performance compares to the norm group.

For example, if you score at the 65th percentile, it means you performed better than 65 percent of the people in the norm group. It does not mean you answered 65 percent of the questions correctly. This is a crucial distinction that many candidates misunderstand.

Norm groups vary depending on the test provider and the employer. Some common norm group categories include:

  • General population norms: Your score is compared against a broad cross-section of working adults. This is the widest benchmark.
  • Graduate norms: Your score is compared against other recent graduates or graduate scheme applicants. Because this group tends to be academically stronger, a given raw score will produce a lower percentile than it would against general population norms.
  • Job-specific norms: Your score is compared against people in a particular profession or industry. A financial analyst norm group, for instance, would include people with strong numerical skills, making the benchmark more competitive.
  • Internal company norms: Some large employers like Deloitte and Unilever build their own norm groups based on the performance of current employees or previous applicant cohorts.

The choice of norm group significantly affects your percentile. A raw score of 28 out of 40 might place you at the 80th percentile against general population norms but only at the 55th percentile against a graduate norm group. This is why you cannot compare percentile scores across different tests or employers without knowing which norm group was used.

💡Your aptitude test score is not about how many questions you got right in isolation. It is about how your performance compares to a specific group of people. Always find out which norm group your employer uses so you can set a realistic target percentile.

Raw Scores vs Scaled Scores vs Standardized Scores

Behind every percentile you see on a results report, there are multiple layers of score transformation. Understanding these layers helps you interpret what your numbers actually mean.

Raw scores are the simplest metric: the number of questions you answered correctly. If a test has 30 questions and you answered 24 correctly, your raw score is 24. However, raw scores are almost never shared with employers because they are not comparable across different versions of the same test. One version might have harder questions than another, so a raw score of 24 on a difficult version could represent stronger ability than a raw score of 27 on an easier version.

Scaled scores solve this problem by placing raw scores onto a common scale that accounts for differences in test difficulty. The scaling process uses statistical techniques like item response theory to ensure that a scaled score of, say, 115 represents the same level of ability regardless of which specific questions you encountered. SHL, Aon, and most other major test providers use scaled scores as an intermediate step before calculating percentiles.

Standardized scores take scaling one step further by expressing your performance in terms of standard deviations from the mean. The most common standardized score formats include:

  • T-scores: A scale with a mean of 50 and a standard deviation of 10. A T-score of 60 means you scored one standard deviation above the average.
  • Stanines: A nine-point scale where a stanine of 5 is average and a stanine of 9 is the highest band. Stanines group candidates into broader categories, which reduces the appearance of small, statistically meaningless differences.
  • Sten scores: A ten-point scale similar to stanines, commonly used in personality and behavioral assessments alongside aptitude tests.

Here is how these different score types compare:

Score Type What It Measures Scale Used By Employers? Shared With Candidates?
Raw Score Number of correct answers Varies by test length Rarely Almost never
Scaled Score Difficulty-adjusted performance Provider-specific (e.g., 0-100, 0-200) Sometimes Occasionally
Percentile Rank relative to norm group 1st to 99th Almost always Sometimes
T-Score Standard deviations from mean Mean 50, SD 10 Sometimes Rarely
Stanine Nine-band classification 1 to 9 Sometimes Sometimes
Sten Score Ten-band classification 1 to 10 For personality tests Sometimes
Pass/Fail Whether you met the cut-off Binary Always when used Usually

The key point is that all of these score types are derived from the same underlying data: your pattern of correct and incorrect answers. They are simply different ways of expressing how well you performed relative to the comparison group. Employers choose whichever format gives them the clearest picture for their decision-making needs.

💡Raw scores tell you almost nothing useful on their own. What matters is how your raw performance translates into a scaled, standardized, or percentile score that can be fairly compared across all candidates.

How Percentile Scores Work in Practice

Percentiles are the most commonly used scoring metric in employer reports, and understanding exactly how they work gives you a significant advantage when interpreting your results and targeting your preparation.

A percentile rank is a number between 1 and 99 that indicates the percentage of the norm group you outperformed. If you are at the 72nd percentile, you scored higher than 72 percent of the people in the norm group and lower than 28 percent. The 50th percentile is the median, meaning you performed exactly in the middle of the distribution.

One of the most important things to understand about percentiles is that they are not evenly spaced in terms of ability. The difference in actual performance between the 50th and 60th percentile is much smaller than the difference between the 85th and 95th percentile. This is because scores tend to cluster around the middle of the distribution (the bell curve effect), so small improvements in raw score can produce large jumps in percentile near the center, while achieving the same percentile jump at the extremes requires much larger improvements in raw performance.

This has a practical implication for your preparation. If you are currently scoring around the 40th to 50th percentile, targeted practice can help you jump to the 60th or 70th percentile relatively quickly because you only need to answer a few more questions correctly. But if you are already at the 80th percentile and need to reach the 90th, you will need to push your accuracy and speed significantly further.

Employers at different competitive levels typically set different percentile thresholds. Entry-level roles at mid-sized companies might require the 30th to 40th percentile. Graduate schemes at firms like Deloitte or Unilever often set cut-offs around the 50th to 60th percentile. Highly competitive programs at organizations like Google, JP Morgan, or the Civil Service Fast Stream may require the 70th percentile or above. These thresholds are not always publicly disclosed, but understanding the general landscape helps you calibrate your preparation effort.

If you want to build the skills and speed needed to reach higher percentile bands, structured practice is essential. Explore aptitude test preparation plans to find targeted practice materials for the specific test format you will face.

Cut-Off Scores: How Employers Filter Candidates

Cut-off scores are the thresholds employers set to determine which candidates advance to the next stage of the hiring process. Understanding how cut-offs work helps you appreciate why scoring "well" is not always enough and why knowing your target employer's standards matters.

There are two main approaches to setting cut-offs:

Fixed cut-offs are predetermined thresholds that remain constant across a recruitment cycle. For example, an employer might decide that all candidates must score at or above the 50th percentile to progress. Everyone who meets the threshold moves forward, regardless of how many candidates that includes. This approach is common among large employers like the Civil Service and Unilever who process thousands of applications and need a consistent, defensible filtering mechanism.

Floating cut-offs adjust based on the quality of the current applicant pool. If an employer needs to fill 20 positions and receives 500 applications, they might set the cut-off wherever it needs to be to produce a manageable shortlist. In a strong applicant year, the effective cut-off might be the 75th percentile. In a weaker year, it might drop to the 55th percentile. JP Morgan and other investment banks often use this approach because their applicant pools vary significantly by year and by office location.

Some employers use multiple cut-offs across different sections of the assessment. A financial services firm might require the 60th percentile on numerical reasoning but only the 40th percentile on verbal reasoning, reflecting the relative importance of each skill for the role. Other employers calculate a composite score across all sections and apply a single cut-off to the combined result.

It is also worth knowing that some employers use aptitude test scores as more than a pass-fail gate. Even after you clear the cut-off, your score may be used to rank candidates, influence interview scheduling priority, or factor into final hiring decisions. At Google, for example, assessment data feeds into a holistic evaluation that considers multiple data points alongside interview performance.

💡Cut-off scores vary widely between employers and can shift from year to year. Do not aim to merely pass the threshold. Aim to score as high as you can, because your score may continue to influence hiring decisions well beyond the initial screening stage.

What Employers Actually See in Your Results Report

When you complete an aptitude test, the employer does not simply receive a single number. They get a detailed report that presents your performance in context. Understanding what this report contains helps you appreciate what employers are evaluating and why certain aspects of your performance matter more than others.

A typical employer results report from a major test provider like SHL, Cubiks/Talogy, or Aon includes the following elements:

Overall percentile rank: This is the headline number that most employers look at first. It gives a quick indication of where you fall relative to other candidates.

Section-level breakdowns: If the assessment included multiple components, such as numerical reasoning, verbal reasoning, and abstract reasoning, the report shows your percentile for each section separately. This allows the employer to see your strengths and weaknesses across different cognitive domains. An employer hiring for a data analyst role might weigh your numerical reasoning score more heavily, while a communications role might prioritize verbal reasoning.

Response patterns and timing data: Most modern test platforms track how long you spent on each question and flag unusual patterns. Extremely fast responses on difficult questions might suggest guessing, while very slow responses on easy questions might indicate test anxiety or distraction. Employers do not usually scrutinize this data question by question, but aggregated timing information is part of the overall picture.

Confidence band or standard error of measurement: No test score is perfectly precise. The report often includes a confidence interval around your percentile to indicate the range within which your true score likely falls. For instance, your reported percentile might be 65 with a 95 percent confidence interval of 58 to 72. This tells the employer that your true ability level is somewhere in that range.

Norm group reference: The report specifies which norm group was used for comparison, allowing the employer to interpret your percentile in context. A percentile of 60 against a graduate norm group signals stronger performance than a percentile of 60 against a general population norm group.

Fit or match indicators: Some providers include an indication of how well your profile matches the requirements of the specific role. This is particularly common with providers like Cappfinity and Pymetrics, whose platforms compare your cognitive and behavioral profile against a model of successful employees in similar roles.

For candidates who want to understand what employers are looking for and how to build a strong profile across all measured dimensions, taking a structured approach to preparation is the most effective strategy. Working through aptitude test questions and answers helps you understand both the content and the scoring logic behind different question types.

How Different Test Providers Score Their Assessments

Each major test provider has its own scoring methodology, and knowing the differences helps you prepare more effectively for the specific test you will face.

SHL uses adaptive testing in many of its assessments, which means the difficulty of each question adjusts based on your previous answers. If you answer a question correctly, the next question becomes harder. If you answer incorrectly, the next question becomes easier. This approach allows SHL to estimate your ability level with fewer questions and greater precision. Your final score reflects not just how many questions you answered correctly but the difficulty level of the questions you got right. SHL reports typically include percentiles, stanines, and sometimes T-scores, all referenced against the employer's chosen norm group.

Cubiks/Talogy uses a fixed-difficulty format for most of its assessments, meaning every candidate sees the same questions at the same difficulty level. The Logiks tests (General, Intermediate, and Advanced) each target a different ability range. Your raw score is converted to a percentile based on the norm group appropriate to the test level. Talogy reports often include section breakdowns and may incorporate personality data from the PAPI questionnaire alongside cognitive scores.

Aon/cut-e uses a distinctive "scales" format consisting of very short subtests, each lasting only a few minutes. The scoring emphasizes speed and accuracy across multiple rapid tasks rather than extended problem-solving. Aon reports include a detailed profile showing your performance across each subtest, and composite scores that combine multiple dimensions into an overall assessment.

Kenexa/IBM focuses heavily on role-specific skill validation. The Prove It assessments measure practical competencies like software proficiency, data entry speed, and job-relevant knowledge alongside traditional aptitude. Scoring is straightforward: your performance is compared to a norm group of people in similar roles, and results are reported as percentiles or pass-fail classifications.

Pymetrics and Arctic Shores use gamified assessments where scoring is fundamentally different from traditional tests. There are no right or wrong answers in the conventional sense. Instead, the system measures behavioral traits like risk tolerance, attention span, learning agility, and decision-making style. Your trait profile is compared against a model built from data on successful employees in the target role. The output is a fit score rather than a traditional percentile.

Understanding which provider your employer uses is a critical first step in preparation. Different formats require different strategies, and practicing with the wrong format can create false confidence. If you are preparing for a specific employer's assessment, check which test provider they use and practice accordingly using resources from assessment-training.com.

How to Improve Your Aptitude Test Score

Now that you understand how scoring works, you can approach your preparation with a clear strategy. Improving your aptitude test score is not about memorizing answers. It is about building the cognitive speed, accuracy, and test-taking skills that translate into higher percentile rankings.

Diagnose your starting point. Take a full-length practice test under realistic timed conditions and calculate your approximate percentile. This gives you a baseline against which to measure improvement. If you are starting at the 40th percentile and need to reach the 60th, you know you need to answer roughly five to eight more questions correctly, depending on the test length.

Focus on your weakest areas. Your percentile is most efficiently improved by fixing your weakest sections rather than polishing your strongest ones. If your numerical reasoning is at the 45th percentile and your verbal reasoning is at the 70th, spending an extra hour on numerical practice will produce a bigger overall improvement than spending that hour on verbal practice.

Build speed through repeated exposure. Most candidates lose points not because they lack the ability to solve the questions but because they run out of time. Repeated practice with timed conditions builds the pattern recognition and calculation shortcuts that allow you to work faster without sacrificing accuracy. For numerical reasoning, practice mental arithmetic, learn to read data tables quickly, and memorize common percentage and fraction conversions. For verbal reasoning, practice distinguishing between what a passage explicitly states and what it merely implies.

Learn the question formats. Every test provider has characteristic question styles. SHL numerical reasoning questions often involve interpreting data from tables and charts. Cubiks Logiks questions test quick arithmetic and logical deduction. Aon scales questions emphasize speed on very short tasks. Practicing with format-specific materials reduces the time you spend figuring out what a question is asking and maximizes the time you spend actually solving it.

Simulate real test conditions. Practice in a quiet room, at a desk, with no distractions. Set a timer and do not pause it. Close all other applications. The goal is to make the real test feel like just another practice session rather than a high-pressure event. Candidates who have done twenty timed practice sessions handle test-day pressure far better than those who have only done casual, untimed review.

Review your mistakes systematically. After each practice session, go through every question you got wrong or guessed on. Identify whether the mistake was due to a knowledge gap, a calculation error, a misreading of the question, or simply running out of time. Each type of mistake requires a different corrective strategy.

If you want to see how different question types work and practice with explained solutions, explore aptitude test tips and strategies for detailed guidance on each reasoning type.

💡Improving your aptitude test score requires targeted, timed practice focused on your weakest areas. A structured preparation plan that simulates real test conditions is far more effective than casual review.

Common Scoring Myths and Misconceptions

Several persistent myths about aptitude test scoring lead candidates to make poor preparation decisions or misinterpret their results. Clearing up these misconceptions helps you approach the process with accurate expectations.

Myth: A higher percentage correct always means a higher percentile. This is false because percentiles are relative to the norm group, not absolute. If you scored 80 percent correct on a very easy test version, your percentile might be lower than someone who scored 70 percent correct on a harder version. Adaptive tests like those from SHL further complicate this because different candidates see different questions at different difficulty levels.

Myth: You are penalized for wrong answers. Most modern aptitude tests use number-right scoring, meaning your score is based solely on the number of correct answers. There is no penalty for incorrect responses. This means you should never leave a question blank. If you are running out of time, make your best guess on remaining questions rather than leaving them unanswered. The exception is a small number of older test formats that explicitly state a guessing penalty, but these are increasingly rare.

Myth: Your percentile score is the same as your percentage score. As discussed above, these are entirely different metrics. Your percentage score tells you how many questions you answered correctly. Your percentile score tells you how you compare to other people. A candidate who answered 60 percent of questions correctly could be at the 30th or the 80th percentile depending on how everyone else performed.

Myth: Aptitude test scores measure intelligence. Aptitude tests measure specific cognitive skills like numerical reasoning, verbal comprehension, and pattern recognition under timed conditions. They are designed to predict job performance, not to provide a comprehensive measure of intelligence. Your score reflects a combination of underlying ability, learned skills, familiarity with the test format, and your state of mind on test day. This is why practice consistently improves scores: you are not becoming more intelligent, but you are getting better at demonstrating your ability under test conditions.

Myth: You cannot improve your aptitude test scores. This is perhaps the most damaging myth because it discourages preparation. Research consistently shows that practice improves aptitude test scores. Candidates who complete structured preparation programs typically improve by 10 to 20 percentile points compared to their initial baseline. The improvement comes from faster processing, better time management, reduced anxiety, and familiarity with question formats.

Myth: All aptitude tests are scored the same way. Different providers use different scoring methodologies, norm groups, and reporting formats. An SHL adaptive test is scored very differently from a Cubiks fixed-difficulty test or a Pymetrics gamified assessment. Preparing as if all tests are identical can leave you unprepared for the specific format and scoring approach you will encounter.

The Role of Aptitude Scores in the Wider Hiring Process

Your aptitude test score does not exist in isolation. It is one component of a multi-stage hiring process, and understanding how it fits into the bigger picture helps you maintain perspective and allocate your preparation time wisely.

For most large employers, the hiring funnel looks something like this. First, your application and CV are screened, either manually or through an applicant tracking system. Next comes the aptitude test, which serves as an objective, standardized filter. Candidates who meet the score threshold advance to further stages, which might include video interviews, assessment centers, group exercises, case studies, or final-round interviews.

At employers like Deloitte, the aptitude test is typically an early-stage filter. They use it to reduce a large applicant pool to a manageable shortlist before investing time in interviews. Your score needs to clear the cut-off, but once you are past that threshold, interview performance becomes the primary differentiator. At Google, assessment data is treated as one input among many in a holistic evaluation, and there is no single score that guarantees or prevents an offer.

The Civil Service Fast Stream takes a different approach, weighting aptitude test scores more heavily in the early rounds and using them alongside situational judgment tests and written exercises to determine who advances to the assessment center. Unilever has pioneered the use of gamified assessments from Pymetrics as an early screening tool, with the AI-driven fit score replacing traditional cut-off percentiles entirely.

JP Morgan uses aptitude assessments alongside coding challenges and technical tests for technology roles, and alongside numerical and verbal reasoning for banking and finance positions. The aptitude score serves as a baseline competency check, and candidates who clear it are evaluated primarily on technical skills and cultural fit in subsequent rounds.

What all of these approaches have in common is that the aptitude test score is a gateway. Falling below the threshold closes the door regardless of how strong the rest of your application might be. But simply clearing the threshold does not guarantee success. You need to perform well across every stage of the process.

For candidates who want to build confidence across all stages, including aptitude tests, interviews, and assessment centers, a comprehensive preparation approach yields the best results. Understanding how tests are scored is the foundation, but pairing that knowledge with consistent, targeted practice is what produces top percentile performance. Explore preparation guidance for specific industries in articles like aptitude tests for graduates to see how scoring applies to your particular career path.

Frequently Asked Questions

What is a good percentile on an aptitude test?

A good percentile depends on the employer and the competitiveness of the role. Many graduate schemes at mid-tier employers set their cut-off around the 50th percentile, meaning you need to outperform at least half of the norm group to advance. However, highly competitive employers like Google, JP Morgan, Deloitte, and the Civil Service Fast Stream often require candidates to reach the 70th percentile or higher. Some niche roles in quantitative finance or data science may require even higher scores. The most effective strategy is not to aim for a specific "good" percentile but to score as high as you possibly can through thorough and structured practice. Every additional percentile point gives you more margin and makes your application more competitive.

Do employers see my raw score?

In the vast majority of cases, employers do not see your raw score. What they receive is a results report from the test provider that shows your percentile rank, often broken down by section (numerical, verbal, abstract reasoning, etc.), along with standardized scores like stanines or T-scores. Raw scores are not shared because they are meaningless without context. A raw score of 22 out of 30 could be excellent or mediocre depending on the difficulty of the questions and the performance of the norm group. Test providers like SHL, Aon, and Cubiks/Talogy convert raw scores into standardized metrics that allow fair comparison across all candidates.

Can I request my aptitude test results?

Policies vary between employers and test providers. Under data protection regulations like GDPR in Europe and similar frameworks in other regions, you generally have the right to request access to personal data held about you, which may include your test scores. Some employers, particularly those using SHL's TalentCentral platform, offer candidate feedback reports automatically. Others only inform you whether you advanced to the next stage. If results are not shared proactively, ask your recruiter directly. Even partial feedback can be valuable for identifying areas to work on if you need to retake the test in a future recruitment cycle.

How is a percentile different from a percentage?

A percentage is a straightforward measure of how many questions you answered correctly out of the total number of questions. If you answered 35 out of 50 questions correctly, your percentage is 70 percent. A percentile, on the other hand, tells you how your performance ranks relative to a group of other test takers. Being at the 70th percentile means you outperformed 70 percent of the norm group, regardless of how many questions you answered correctly. The two numbers can diverge significantly. On a difficult test where most candidates score below 60 percent, answering 65 percent correctly might place you at the 85th percentile. On an easier test where most candidates score above 80 percent, answering 75 percent correctly might only place you at the 30th percentile. Employers use percentiles rather than percentages because percentiles account for test difficulty and provide a fair basis for comparing candidates.

What happens if I score below the cut-off?

If your score falls below the employer's cut-off threshold, your application is typically screened out and you will not advance to the next stage of the hiring process. The employer may or may not inform you that the aptitude test was the reason. Some employers allow you to retake the assessment after a waiting period, which is usually six to twelve months. Others only permit one attempt per recruitment cycle, meaning you would need to wait until the next intake to reapply. In either scenario, the best approach is to use the intervening time to practice systematically. Identify the sections where you scored lowest and work on building speed and accuracy in those areas. A structured preparation plan can help you improve by 10 to 20 percentile points before your next attempt.

Are aptitude test scores comparable across different test providers?

No, aptitude test scores are not directly comparable across different providers. Each provider uses its own question bank, scoring algorithm, norm group, and standardization method. A percentile of 75 on an SHL Verify test represents your performance relative to SHL's specific norm group, which may have a different composition and size than the norm group used by Cubiks/Talogy or Aon. The difficulty calibration, adaptive versus fixed-format design, and timing structure also differ between providers, making cross-provider comparisons unreliable. Employers understand this and evaluate your score only within the context of the specific test they administered. This is also why it is important to practice with materials that match the format and provider you will actually encounter rather than assuming that any generic aptitude practice will transfer equally well to all tests.

Start Improving Your Aptitude Test Score Today

Understanding how aptitude tests are scored gives you a real strategic advantage. You now know that your score is not just about getting questions right but about where your performance falls relative to a specific group of candidates. You understand the difference between raw scores, scaled scores, and percentiles. You know how cut-offs work and what employers actually see in your results report.

The next step is to put that knowledge into practice. Candidates who combine scoring knowledge with structured, timed practice consistently outperform those who go in without preparation. Whether you are targeting a graduate scheme at Deloitte, a technology role at Google, an analyst position at JP Morgan, or a Civil Service Fast Stream place, the scoring principles are the same: build speed, improve accuracy, and push your percentile as high as possible.

Get started with the complete test package at assessment-training.com to access practice tests covering SHL, Cubiks/Talogy, Kenexa, Aon, and other major providers. Build the skills and confidence you need to score at your best.