You're a two-person talent team at a 40-person company. A hiring manager pings you on Slack asking if you can "just add an English test" to the funnel for a customer support position.
You open a tab, type "English literacy test for hiring," and land on a vendor page selling Versant scores for $30 a pop. Twenty minutes per candidate. Instant results. It feels easy. You almost click buy.
Then you remember the last time your legal counsel saw a blanket pre-employment test. The call didn't go well. You close the tab.
Here's the thing. If you're reaching for an off-the-shelf English literacy test, what you actually want is evidence the candidate can do the position in English. That's a job-relevance question, not a reading comprehension question. And the difference between those two things is the difference between a defensible screen and an EEOC complaint.
This post walks through the legal risk, what to actually measure, and six alternatives you can run this week.
Disclaimer: This post is for informational purposes only and is not legal advice. Employment law varies by jurisdiction and situation. Before changing any pre-employment screen, consult a qualified employment attorney about your specific positions, candidate pool, and applicable regulations.
The legal problem with generic English tests
You're right to be nervous about this. A standalone English literacy test, dropped at the top of the funnel for every candidate, is one of the riskier screening choices you can make.
Title VII of the Civil Rights Act prohibits employment practices that have a disparate impact on protected groups, including national origin and race. A generic English test can screen out otherwise qualified candidates whose first language isn't English, even when the position doesn't require that level of English. That pattern is what disparate-impact case law was built to catch.
The legal standard is specific. Under the Uniform Guidelines on Employee Selection Procedures (29 CFR Part 1607), any employment test that produces disparate impact has to be job-related for the position and consistent with business necessity. The EEOC uses a "four-fifths rule" as a practical threshold: if the pass rate for a protected group is less than 80% of the pass rate for the highest-scoring group, the test is presumed to have adverse impact unless the employer can demonstrate job-relatedness.
A customer support position that handles written English chat all day has a very different English requirement than a warehouse position that mostly needs spoken safety comprehension. A blanket literacy test treats them the same. A job-related screen doesn't.
This post isn't legal advice. Before you change any pre-employment screen, talk to your employment counsel about your specific positions, candidate pool, and the jurisdictions you hire in.
What I can tell you is that "job-related and documented" holds up a lot better than "we bought a generic test off the shelf." Our guide to hiring assessments covers screen design more broadly. For a parallel analysis on personality tests, see our legal risk audit for personality assessments.
What off-the-shelf English tests actually measure
Most of the English proficiency tests you'll find, Versant (Pearson), Duolingo English Test, LinguaSkill (Cambridge), Criteria Corp's CLPT, are built around a general proficiency framework called CEFR (Common European Framework of Reference for Languages). CEFR maps English ability to six levels, from A1 (beginner) to C2 (proficient).
These tests are well-validated for what they measure: general reading, writing, speaking, and listening ability on a standardized scale. The problem is what they don't measure. A CEFR score tells you a candidate is "B2 upper-intermediate." It doesn't tell you whether they can handle an angry customer in a live chat window, or write a concise incident report under time pressure, or follow a dispatcher's instructions on a noisy warehouse floor.
General proficiency and job-relevant communication are different things. A candidate at B1 who's spent two years handling English-language support tickets might outperform a C1 candidate who's never worked in a customer-facing role. The standardized test can't see that difference. A job-relevant screen can.
What you actually need to measure
Step back from the word "literacy" for a minute. English has four skill domains, and every position weights them differently:
- Reading: Can the candidate parse written instructions, emails, or documentation at the speed the position requires? A QA analyst reading bug reports has different reading demands than a receptionist scanning appointment confirmations.
- Writing: Can the candidate produce the written artifacts the position requires? A support rep drafting ticket responses needs different writing skills than a marketing coordinator writing product updates.
- Speaking: Can the candidate communicate verbally in the position's actual context? A phone-based account manager needs fluent spoken English. A data entry clerk working independently may not.
- Listening: Can the candidate understand spoken English in the position's environment? A dispatcher taking calls has different listening demands than a warehouse worker following safety briefings.
Write down which of these four domains the position actually requires, and at what level. One paragraph per position, signed off by the hiring manager. That paragraph becomes your screening spec. Everything else follows from it.
For a customer support rep, maybe it's "can read a customer's chat message, understand the issue, and write a clear reply in under two minutes." For a field technician, maybe it's "can follow a written safety procedure and answer a dispatcher's questions on the phone."
Those are concrete. They're testable. And they're defensible. Notice how none of them sound like a reading comprehension exam. See our post on candidate assessment for how this fits a bigger evaluation plan.
Six job-related alternatives to an English literacy test
Pick one or two. Don't stack all six. The goal is a screen that matches the position's real English demands, not a gauntlet. Use this table to find the right fit, then read the details below.
1. Structured work samples in English
A work sample is a short task that mirrors the position. For a support rep, it might be three sample customer messages and 15 minutes to draft replies. For a technical writer, it might be a 200-word explainer on a feature you sell. You score it against a rubric the hiring manager builds in advance, and every candidate gets the same prompt and the same rubric.
Work samples are among the most predictive screening methods available. The Schmidt-Hunter meta-analysis of selection methods found that work sample tests have a validity coefficient of .54, higher than job interviews (.51) and far above resume screening (.18). They're naturally job-related because you built them from the position itself. For a deeper cut on why, see our list of the best hiring assessments.
2. One-way video interviews with role-specific prompts
A one-way interview lets candidates record answers to prompts on their own time. You write prompts that pull from real position scenarios, and the candidate responds on camera. You watch when you have time, and you can review the same answer twice if you need to.
For English comprehension and clarity, this is a clean signal. You hear how the candidate speaks in the position's actual context, not a context invented by a test vendor. A good prompt for a support position sounds like "a customer just said their package arrived broken and they want a refund today. Walk us through how you'd reply." That's the position.
This also works better from the candidate's side. Recording a response to a job-relevant prompt on your own schedule is less intimidating than sitting for a standardized language exam, especially for candidates who communicate well in practice but freeze during formal testing. You get a more authentic signal and a better candidate experience.
Truffle is candidate screening software that combines resume screening, one-way video interviews, and talent assessments. For English screening specifically, you write prompts pulled from the position.
Candidates record answers on camera. Truffle's AI Summaries pull the key points from each response, and Candidate Shorts give you a 30-second highlight clip per candidate so you can scan 40 responses in 20 minutes.
AI Match surfaces strong candidates based on your criteria, and you and your hiring manager make the call. The AI surfaces. You decide.
3. Situational judgment tests in context
A situational judgment test presents a realistic scenario and asks the candidate to pick or describe the best response. For English in context, you can write scenarios around the position's actual communication demands. A scenario for a restaurant host might describe a guest with a reservation complaint, and the candidate picks from four possible responses.
Because the scenarios come from the position, the English load matches the position. You're not asking candidates to parse academic passages. You're asking if they can handle the communication the position actually involves.
4. Role-play or customer-call simulations
For any position that talks to customers, a role-play is hard to beat. A short mock call, maybe five minutes, gives you a real sample of how the candidate handles the pressure and vocabulary of the position. You can run this live or as a one-way video with a recorded customer prompt.
The value isn't just language. You also see how they handle uncertainty, tone, and the back-and-forth of a real conversation. That's much closer to the position than a multiple-choice test will ever get.
5. Written task scored against job-specific criteria
Some positions are mostly written communication. For those, a written task scored against a rubric built around the position is the most defensible screen you can use. Pick a real artifact the position produces, shorten it, and grade against a rubric the hiring manager signs off on.
Examples. A support position gets a ticket draft. A marketing coordinator gets a 150-word product update. A QA analyst gets a bug report. Each one maps directly to position output, so the language load matches what the position needs.
6. Paired-task collaboration with a teammate
For senior positions or positions where team communication matters most, have the candidate work through a short task with a current team member on a video call. You're watching the collaboration, not grading a test. Can they ask clarifying questions? Do they understand what the teammate is saying? Can they push back respectfully when they disagree?
This is high-signal and low-volume, so save it for later rounds. But as a late-stage check on working English, it's honest and fair. It's also a natural fit alongside blind hiring practices and inclusive job descriptions earlier in the funnel.
How to switch from a literacy test to a job-relevant screen
If you're on a generic English test and want to move to something defensible, here's a four-step plan that doesn't require rebuilding your pipeline.
- Talk to your employment counsel first. Describe your current screen, your candidate pool, and the positions you use it on. Get their read before you change anything.
- Write down each position's real English demands. One paragraph per position, specifying which skill domains (reading, writing, speaking, listening) the position requires and at what level. Have the hiring manager sign off.
- Pick one alternative from the six above. Don't stack. One well-designed screen beats three generic ones. Use the decision table to match the alternative to the position type.
- Document everything. Keep the position language spec, the rubric, and records of how you scored each candidate. If you ever need to defend the screen under the Uniform Guidelines, that paper trail is your best asset.
That's it. You're not changing your ATS or buying a new stack. You're swapping one step for a better one. Our guide to pre-employment assessment software has more on sequencing, and our recruitment assessment tools roundup covers the broader landscape.
Frequently asked questions about English language assessments for hiring
Is an English literacy test legal for hiring?
It depends on how you use it. A generic English test applied as a blanket screen to every candidate is legally risky under Title VII because it can create disparate impact based on national origin. The Uniform Guidelines on Employee Selection Procedures (29 CFR Part 1607) require that any test with adverse impact be validated as job-related and consistent with business necessity. A job-related English screen, tied to the communication demands of a specific position and documented as such, is a much safer approach. Talk to your employment counsel before you make any changes.
What's the difference between an English proficiency test for employment and a job-relevant screen?
A proficiency test measures general English ability on a standardized scale like CEFR (A1 through C2). A job-relevant screen measures whether the candidate can handle the English demands of a specific position. The first treats every position the same. The second matches the screen to the position, which is both more predictive and more defensible. A candidate who scores B1 on a CEFR test might be perfectly effective in a position that requires basic written English and strong verbal skills.
How do I document that an English screen is job-related?
Start with a written description of the position's language demands, specifying which of the four skill domains (reading, writing, speaking, listening) are required and at what level. Have the hiring manager sign off. Build your screen directly from that description. Keep a rubric and use it consistently for every candidate. Save records of how you scored each response. This documentation is what the Uniform Guidelines call "evidence of validity," and it's your best defense if the screen is ever challenged.
Can I use a workplace English test for customer-facing positions?
You can, if it's built around the position. A customer service rep who handles written chat has different English demands than one who takes phone calls, and your screen should reflect that. Pull prompts, scenarios, and rubrics from the position itself rather than from a generic vendor template, and you'll end up with a screen that's both fairer and more predictive.
What is CEFR and does it matter for hiring?
CEFR (Common European Framework of Reference for Languages) is a standardized scale for measuring language proficiency, from A1 (beginner) to C2 (proficient). Most off-the-shelf English proficiency tests, including Versant, Duolingo English Test, and LinguaSkill, map results to CEFR levels. The framework is useful as a shared vocabulary ("this position requires B2-level writing"), but a CEFR score alone doesn't tell you whether a candidate can do the position. Use CEFR as a reference point for setting expectations, not as your entire screen.
Do I need to validate my English test under the Uniform Guidelines?
If your test produces disparate impact on a protected group (which generic English tests often do, by national origin), the Uniform Guidelines require evidence that the test is job-related and consistent with business necessity. Validation can be criterion-related, content-based, or construct-based. The simplest path for most employers is content validity: show that the test directly samples the skills the position requires. A work sample or role-play built from actual position tasks is inherently content-valid. A generic reading comprehension exam is much harder to defend.
What if my team works in multiple languages?
If the position operates in a multilingual environment, your screen should reflect that reality. Define which tasks require English, which require another language, and whether the position needs to switch between them. A blanket English-only test doesn't match a position where half the team communicates in Spanish and the other half in English. Screen for the actual communication pattern, not an idealized monolingual one. Our guide to hiring in South Africa covers multi-language hiring in more detail.




