Reading
Part 6 - Cross Matching
You are going to read a series of texts. For questions 1-4, choose the correct text. Texts can be chosen more than once.
Cross-Text Perspectives
Four experts/writers give their perspectives on producing a C1 Cambridge-style Reading & Use of English cross-text multiple-matching exercise.
Writer A
If you want a task that *genuinely* mimics Cambridge C1, you can’t just sprinkle in a few fancy words and call it a day. What matters is the calibration: each text must be compact yet conceptually loaded, with stance signalled through hedging, concessive clauses and carefully chosen evaluative lexis. Nor should the questions reward crude keyword-spotting; the whole point is to force candidates to track *attitude* and *reasoning*. That said, authenticity is not synonymous with opacity. A well-designed set is demanding because it is nuanced, not because it is wilfully obscure. I’m also wary of over-engineering ‘trick’ distractors: the best items mislead incidentally, through plausible overlap, rather than through gimmickry.
Writer B
Accuracy to the exam is, in practice, a matter of replicating its *constraints*: four voices, a shared topic, and questions that require cross-referencing positions rather than extracting facts. The temptation is to make every sentence baroque; yet Cambridge’s difficulty often lies in how ordinary phrasing masks a subtle shift in commitment—*may well*, *hardly*, *for all that*. I’d argue that the most faithful imitation includes a disciplined register and a clear rhetorical spine in each text. And yes, you should include anti-keyword traps: if a question mentions ‘authenticity’, let a non-answer text use that very word, while the correct answer expresses the idea via paraphrase. Otherwise, candidates can game the task.
Writer C
I’m less convinced that ‘mimicking Cambridge’ should be the overriding aim. Teaching materials that fetishise exam likeness can end up narrowing reading into a scavenger hunt for stance markers. What I’d prioritise is transferability: texts that resemble real-world commentary—editorials, reviews, think-pieces—so that the skill is robust beyond the test. Ironically, the more you chase ‘exam authenticity’, the more you risk producing contrived mini-essays that no one would ever write. Of course, paraphrase-heavy questions are essential, but I’d rather they illuminate meaning than function as booby traps. A candidate should feel challenged, not ambushed.
Writer D
The uncomfortable truth is that most home-made cross-matching tasks fail because they don’t control for *comparability*. Writers drift onto adjacent subtopics, so the questions become guesswork. Cambridge avoids this by making each voice address the same handful of dimensions—value, feasibility, consequences—while disagreeing in fine-grained ways. Where I part company with some item-writers is on ‘fairness’: a task can be fair and still be ruthless. If anything, the exam’s hallmark is that it punishes superficial reading. So I’m in favour of deliberate distractor engineering, provided it’s principled: lexical overlap in the wrong place, and the correct answer expressed through a higher-level reformulation.
1. Which writer shares Writer A’s view that a high-fidelity task should test nuanced stance rather than rely on gratuitous obscurity?
2. Which writer has a different view from the others on whether reproducing Cambridge exam-likeness ought to be the primary objective?
3. Which writer expresses a similar view to Writer C regarding the risk that exam-style materials can feel contrived rather than resembling genuine discourse?
4. Which writer most directly contrasts with Writer A on the desirability of intentionally engineered distractors and ‘trap’ design?
Instructions
You are going to read a series of texts. For questions 1-4, choose the correct text. Texts can be chosen more than once.
Exercise Details
Author
Sabina Chialda
@sabina-e1bff3
User Prompt
"Produce a C1 Reading & Use of English cross-text multiple-matching exercise that mimics the Cambridge English exam as accurately as possible."
Created on:
Apr 1, 2026
Found an issue? Let us know.
