diff --git "a/txt/2204.03592.txt" "b/txt/2204.03592.txt" deleted file mode 100644--- "a/txt/2204.03592.txt" +++ /dev/null @@ -1,1569 +0,0 @@ -Testing the limits of natural language models for -predicting human language judgments -Tal Golan1,2∗†, Matthew Siegelman3∗, -Nikolaus Kriegeskorte1,3,4,5, Christopher Baldassano3 -1Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY , USA -2Department of Cognitive and Brain Sciences, Ben-Gurion University of the Negev, Be’er-Sheva, Israel -3Department of Psychology, Columbia University, New York, NY , USA -4Department of Neuroscience, Columbia University, New York, NY , USA -5Department of Electrical Engineering, Columbia University, New York, NY , USA -∗The first two authors contributed equally to this work. -†To whom correspondence should be addressed; E-mail: golan.neuro@bgu.ac.il -Neural network language models appear to be increasingly aligned with how hu- -mans process and generate language, but identifying their weaknesses through ad- -versarial examples is challenging due to the discrete nature of language and the -complexity of human language perception. We bypass these limitations by turning -the models against each other. We generate controversial sentence pairs for which -two language models disagree about which sentence is more likely to occur. Con- -sidering nine language models (including n-gram, recurrent neural networks, and -transformers), we created hundreds of controversial sentence pairs through syn- -thetic optimization or by selecting sentences from a corpus. Controversial sentence -pairs proved highly effective at revealing model failures and identifying models that -aligned most closely with human judgments of which sentence is more likely. The -most human-consistent model tested was GPT-2, although experiments also revealed -significant shortcomings of its alignment with human perception. -Keywords— Language Models, Human Acceptability Judgments, Controversial Stimuli, Adversarial Attacks -in NLP -1 Introduction -Neural network language models are not only key tools in natural language processing (NLP) but are also draw- -ing an increasing scientific interest as potential models of human language-processing. Ranging from recurrent -neural networks [1, 2] to transformers [3–7], each of these language models (explicitly or implicitly) defines -a probability distribution over strings of words, predicting which sequences are likely to occur in natural lan- -guage. There is substantial evidence from measures such as reading times [8], functional MRI [9], scalp EEG -1arXiv:2204.03592v3 [cs.CL] 12 Sep 2023[10], and intracranial ECoG [11] that humans are sensitive to the relative probabilities of words and sentences as -captured by language models, even among sentences that are grammatically correct and semantically meaning- -ful. Furthermore, model-derived sentence probabilities can also predict human graded acceptability judgments -[12, 13]. These successes, however, have not yet addressed two central questions of interest: (1) Which of the -models is best-aligned with human language processing? (2) How close is the best-aligned model to the goal -of fully capturing human judgments? -A predominant approach for evaluating language models is to use a set of standardized benchmarks such -as those in the General Language Understanding Evaluation (GLUE) [14], or its successor, SuperGLUE [15]. -Though instrumental in evaluating the utility of language models for downstream NLP tasks, these benchmarks -prove insufficient for comparing such models as candidate explanations of human language-processing. Many -components of these benchmarks do not aim to measure human alignment but rather the usefulness of the mod- -els’ language representation when tuned to a specific downstream task. Some benchmarks challenge language -models more directly by comparing the probabilities they assign to grammatical and ungrammatical sentences -(e.g., BLiMP [16]). However, since such benchmarks are driven by theoretical linguistic considerations, they -might fail to detect novel, unexpected ways in which language models may diverge from human language un- -derstanding. Lastly, an additional practical concern is that the rapid pace of NLP research has led to quick -saturation of these kinds of static benchmarks, making it difficult to distinguish between models [17]. -One proposed solution to these issues is the use of dynamic human-in-the-loop benchmarks in which peo- -ple actively stress-test models with an evolving set of tests. However, this approach faces the major obstacle -that “finding interesting examples is rapidly becoming a less trivial task” [17]. We propose to complement -human-curated benchmarks with model-driven evaluation. Guided by model predictions rather than experi- -menter intuitions, we would like to identify particularly informative test sentences, where different models -make divergent predictions. This approach of running experiments mathematically optimized to “put in jeop- -ardy” particular models belongs to a long-standing scientific philosophy of design optimization [18]. We can -find these critical sentences in large corpora of natural language or synthesize novel test sentences that reveal -how different models generalize beyond their training distributions. -We propose here a systematic, model-driven approach for comparing language models in terms of their -consistency with human judgments. We generate controversial sentence pairs : pairs of sentences designed -such that two language models strongly disagree about which sentence is more likely to occur. In each of these -sentence pairs, one model assigns a higher probability to the first sentence than the second sentence, while the -other model prefers the second sentence to the first. We then collect human judgments of which sentence in -each pair is more probable to settle this dispute between the two models. -This approach builds on previous work on controversial images for models of visual classification [19]. -That work relied on absolute judgments of a single stimulus, which are appropriate for classification responses. -However, asking the participants to rate each sentence’s probability on an absolute scale is complicated by -between-trial context effects common in magnitude estimation tasks [20–22], which have been shown to impact -judgments like acceptability [23]. A binary forced-choice behavioral task presenting the participants with a -choice between two sentences in each trial, the approach we used here, minimizes the role of between-trial -context effects by setting an explicit local context within each trial. Such an approach has been previously -used for measuring sentence acceptability [24] and provides substantially more statistical power compared to -designs in which acceptability ratings are provided for single sentences [25]. -Our experiments demonstrate that (1) it is possible to procedurally generate controversial sentence pairs for -all common classes of language models, either by selecting pairs of sentences from a corpus or by iteratively -modifying natural sentences to yield controversial predictions; (2) the resulting controversial sentence pairs -enable efficient model comparison between models that otherwise are seemingly equivalent in their human -2consistency; and (3) all current NLP model classes incorrectly assign high probability to some non-natural -sentences (one can modify a natural sentence such that its model probability does not decrease but human -observers reject the sentence as unnatural). This framework for model comparison and model testing can give -us new insight into the classes of models that best align with human language perception and suggest directions -for future model development. -2 Results -We acquired judgments from 100 native English speakers tested online. In each experimental trial, the partic- -ipants were asked to judge which of two sentences they would be “more likely to encounter in the world, as -either speech or written text”, and provided a rating of their confidence in their answer on a 3-point scale (see -Extended Data Fig. 1 for a trial example). The experiment was designed to compare nine different language -models (Supplementary Section 6.1): probability models based on corpus frequencies of 2-word and 3-word -sequences (2-grams and 3-grams) and a range of neural network models comprising a recurrent neural network -(RNN), a long short-term memory network (LSTM), and five transformer models (BERT, RoBERTa, XLM, -ELECTRA, and GPT-2). -2.1 Efficient model comparison using natural controversial pairs -As a baseline, we randomly sampled and paired 8-word sentences from a corpus of Reddit comments. How- -ever, as shown in Fig. 1a, these sentences fail to uncover meaningful differences between the models. For -each sentence pair, all models tend to prefer the same sentence (Extended Data Fig. 2), and therefore perform -similarly in predicting human preference ratings (see Supplementary Section 7.1). -Instead, we can use an optimization procedure (Supplementary Section 6.2) to search for controversial -sentence pairs, in which one language model assigns a high probability (above the median probability for -natural sentences) only to sentence 1 and a second language model assigns a high probability only to sentence -2; see examples in Table 1. Measuring each model’s accuracy in predicting human choices for sentence pairs -in which it was one of the two targeted models indicated many significant differences in terms of model- -human alignment (Fig. 1b), with GPT-2 and RoBERTa showing the best human consistency and 2-gram the -worst. We can also compare each model pair separately (using only the stimuli targeting that model pair), -yielding a similar pattern of pairwise dominance (Extended Data Fig. 3a). All models except GPT-2, RoBERTa, -and ELECTRA performed significantly below our lower bound on the noise ceiling (the accuracy obtained -by predicting each participant’s responses from the other participants’ responses), indicating a misalignment -between these models’ predictions and human judgments which was only revealed when using controversial -sentence pairs. -30 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -0 50 100 -GPT-2 -p(sentence) percentile020406080100RoBERTa -p(sentence) percentile -0 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -0 50 100 -GPT-2 -p(sentence) percentile020406080100RoBERTa -p(sentence) percentile -aRandomly sampled natural-sentence pairs -bControversial natural-sentence pairsFigure 1: Model comparison using natural sentences. (a) (Left) Percentile-transformed sentence probabili- -ties for GPT-2 and RoBERTa (defined relative to all sentences used in the experiment) for randomly-sampled -pairs of natural sentences. Each pair of connected dots depicts one sentence pair. The two models are highly -congruent in their rankings of sentences within a pair (lines have upward slope). (Right) Accuracy of model -predictions of human choices, measured as the proportion of trials in which the same sentence was preferred -by both the model and the human participant. Each dot depicts the prediction accuracy of one candidate model -averaged across a group of 10 participants presented with a unique set of trials. The colored bars depict grand- -averages across all 100 participants. The gray bar is the noise ceiling whose left and right edges are lower -and upper bounds on the grand-average performance an ideal model would achieve (based on the consistency -across human subjects). There were no significant differences in model performance on the randomly sam- -pled natural sentences. (b)(Left) Controversial natural-sentence pairs were selected such that the models’ -sentence probability ranks were incongruent (lines have downward slope). (Right) Controversial sentence pairs -enable efficient model comparison, revealing that BERT, XLM, LSTM, RNN and the n-gram models perform -significantly below the noise ceiling (asterisks indicate significance—two-sided Wilcoxon signed-rank test, -controlling the false discovery rate for nine comparisons at q<.05). On the right of the plot, each closed circle -indicates a model significantly dominating alternative models indicated by open circles (two-sided Wilcoxon -signed-rank test, controlling the false discovery rate for all 36 model pairs at q<.05). GPT-2 outperforms all -models except RoBERTA at predicting human judgments. -4sentence log probability (model 1) log probability (model 2) # human choices -n1: Rust is generally caused by salt and sand. logp(n1|GPT-2 ) =−50.72 log p(n1|ELECTRA ) =−38.54 10 -n2: Where is Vernon Roche when you need him. logp(n2|GPT-2 ) =−32.26 logp(n2|ELECTRA ) =−58.26 0 -n1: Excellent draw and an overall great smoking experience. logp(n1|RoBERTa ) =−67.78 log p(n1|GPT-2 ) =−36.76 10 -n2: I should be higher and tied to inflation. logp(n2|RoBERTa ) =−54.61 logp(n2|GPT-2 ) =−50.31 0 -n1: You may try and ask on their forum. logp(n1|ELECTRA ) =−51.44 log p(n1|LSTM ) =−44.24 10 -n2: I love how they look like octopus tentacles. logp(n2|ELECTRA ) =−35.51 logp(n2|LSTM ) =−66.66 0 -n1: Grow up and quit whining about minor inconveniences. logp(n1|BERT ) =−82.74 log p(n1|GPT-2 ) =−35.66 10 -n2: The extra a is the correct Sanskrit pronunciation. logp(n2|BERT ) =−51.06 logp(n2|GPT-2 ) =−51.10 0 -n1: I like my password manager for this reason. logp(n1|XLM ) =−68.93 log p(n1|RoBERTa ) =−49.61 10 -n2: Kind of like clan of the cave bear. logp(n2|XLM ) =−44.24 logp(n2|RoBERTa ) =−67.00 0 -n1: We have raised a Generation of Computer geeks. logp(n1|LSTM ) =−66.41 log p(n1|ELECTRA ) =−36.57 10 -n2: I mean when the refs are being sketchy. logp(n2|LSTM ) =−42.04 logp(n2|ELECTRA ) =−52.28 0 -n1: This is getting ridiculous and ruining the hobby. logp(n1|RNN) =−100.65 log p(n1|LSTM ) =−43.50 10 -n2: I think the boys and invincible are better. logp(n2|RNN) =−45.16 logp(n2|LSTM ) =−59.00 0 -n1: Then attach them with the supplied wood screws. logp(n1|3-gram ) =−119.09 log p(n1|GPT-2 ) =−34.84 10 -n2: Sounds like you were used both a dog. logp(n2|3-gram ) =−92.07 logp(n2|GPT-2 ) =−52.84 0 -n1: Cream cheese with ham and onions on crackers. logp(n1|2-gram ) =−131.99 log p(n1|RoBERTa ) =−54.62 10 -n2: I may have to parallel process that drinking. logp(n2|2-gram ) =−109.46 logp(n2|RoBERTa ) =−70.69 0 -Table 1: Examples of controversial natural-sentence pairs that maximally contributed to each model’s -prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which -the model failed severely. In each case, the failing model 1 prefers sentence n2(higher log probability bolded), -while the model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair -prefer sentence n1. (When more than one sentence pair induced an equal maximal error in a model, the example -included in the table was chosen at random.) -2.2 Greater model disentanglement with synthetic sentence pairs -Selecting controversial natural-sentence pairs may provide greater power than randomly sampling natural- -sentence pairs, but this search procedure considers a very limited part of the space of possible sentence pairs. -Instead, we can iteratively replace words in a natural sentence to drive different models to make opposing -predictions, forming synthetic controversial sentences that may lay outside any natural language corpora, as il- -lustrated in Fig. 2 (see Methods, “Generating synthetic controversial sentence pairs” for full details). Examples -of controversial synthetic-sentence pairs that maximally contributed to the models’ prediction error appear in -Table 2. -We evaluated how well each model predicted the human sentence choices in all of the controversial synthetic- -sentence pairs in which the model was one of the two models targeted (Fig. 3a). This evaluation of model- -human alignment resulted in an even greater separation between the models’ prediction accuracies than was -obtained when using controversial natural-sentence pairs, pushing the weaker models (RNN, 3-gram, and 2- -gram) far below the 50% chance accuracy level. GPT-2, RoBERTa, and ELECTRA were found to be signifi- -cantly more accurate than the alternative models (BERT, XLM, LSTM, RNN, 3-gram, and 2-gram) in predicting -the human responses to these trials (with similar results when comparing model pair separately, see Extended -Data Fig. 3b). All of the models except for GPT-2 were found to be significantly below the lower bound on the -noise ceiling, demonstrating misalignment with human judgments. -50 20 40 60 80 100 -GPT-2 -p(sentence) percentile020406080100ELECTRA -p(sentence) percentileNothing has -a world of -excitement -and joys. -Diddy has a wealth of experience -with grappling.Luke has a ton of experience -with winning. -a -0 20 40 60 80 100 -RoBERTa -p(sentence) percentile0204060801003-gram -p(sentence) percentileYou have to realize -is that noise again. -I wait to see how -it shakes out.I need to see how -this played out. -bFigure 2: Synthesizing controversial sentence pairs. The small open dots denote 500 randomly sampled -natural sentences. The big open dot denotes the natural sentence used for initializing the controversial sentence -optimization, and the closed dots are the resulting synthetic sentences. (a)In this example, we start with the -randomly sampled natural sentence “Luke has a ton of experience with winning”. If we adjust this sentence to -minimize its probability according to GPT-2 (while keeping the sentence at least as likely as the natural sentence -according to ELECTRA), we obtain the synthetic sentence “Nothing has a world of excitement and joys”. By -repeating this procedure while switching the roles of the models, we generate the synthetic sentence “Diddy -has a wealth of experience with grappling”, which decreases ELECTRA’s probability while slightly increasing -GPT-2’s. (b)In this example, we start with the randomly sampled natural sentence “I need to see how this -played out”. If we adjust this sentence to minimize its probability according to RoBERTa (while keeping the -sentence at least as likely as the natural sentence according to 3-gram), we obtain the synthetic sentence “You -have to realize is that noise again”. If we instead decrease only 3-gram’s probability, we generate the synthetic -sentence “I wait to see how it shakes out”. -2.3 Pairs of natural and synthetic sentences uncover blindspots -Last, we considered trials in which the participants were asked to choose between a natural sentence and one of -the synthetic sentences which was generated from that natural sentence. If the language model is fully aligned -with human judgments, we would expect humans to agree with the model, and select the synthetic sentence -at least as much as the natural sentence. In reality, human participants showed a systematic preference for the -natural sentences over their synthetic counterparts (Fig. 3b), even when the synthetic sentences were formed -such that the stronger models (i.e., GPT-2, RoBERTA, or ELECTRA) favored them over the natural sentences; -see Extended Data Table 1 for examples. Evaluating natural sentence preference separately for each model- -pairing (Extended Data Fig. 4), we find that these imperfections can be uncovered even when pairing a strong -model with a relatively weak model (such that the strong model “accepts” the synthetic sentence and the weak -model rejects it). -60 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -0 50 100 -GPT-2 -p(sentence) percentile020406080100RoBERTa -p(sentence) percentile -0 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -0 50 100 -GPT-2 -p(sentence) percentile020406080100RoBERTa -p(sentence) percentile -aSynthetic controversial sentence pairs -bSynthetic vs. natural sentences -natural sentence synthetic sentenceFigure 3: Model comparison using synthetic sentences. (a) (Left) Percentile-transformed sentence probabil- -ities for GPT-2 and RoBERTa for controversial synthetic-sentence pairs. Each pair of connected dots depict -one sentence pair. (Right) Model prediction accuracy, measured as the proportion of trials in which the same -sentence was preferred by both the model and the human participant. GPT-2, RoBERTa and ELECTRA sig- -nificantly outperformed the other models (two-sided Wilcoxon signed-rank test, controlling the false discovery -rate for all 36 model comparisons at q<.05). All of the models except for GPT-2 were found to perform below -the noise ceiling (gray) of predicting each participant’s choices from the majority votes of the other participants -(asterisks indicate significance—two-sided Wilcoxon signed-rank test, controlling the false discovery rate for -nine comparisons at q<.05). (b)(Left) Each connected triplet of dots depicts a natural sentence and its derived -synthetic sentences, optimized to decrease the probability only under GPT-2 (left dots in a triplet) or only under -RoBERTa (bottom dots in a triplet). (Right) Each model was evaluated across all of the synthetic-natural sen- -tence pairs for which it was targeted to keep the synthetic sentence at least as probable as the natural sentence -(see Extended Data Fig. 6 for the complementary data binning). This evaluation yielded a below-chance pre- -diction accuracy for all of the models, which was also significantly below the lower bound on the noise ceiling. -This indicates that, although the models assessed that these synthetic sentences were at least as probable as the -original natural sentence, humans disagreed and showed a systematic preference for the natural sentence. See -Fig. 1’s caption for details on the visualization conventions used in this figure. -7sentence log probability (model 1) log probability (model 2) # human choices -s1: You can reach his stories on an instant. logp(s1|GPT-2 ) =−64.92 log p(s1|RoBERTa ) =−59.98 10 -s2: Anybody can behead a rattles an an antelope. logp(s2|GPT-2 ) =−40.45 logp(s2|RoBERTa ) =−90.87 0 -s1: However they will still compare you to others. logp(s1|RoBERTa ) =−53.40 log p(s1|GPT-2 ) =−31.59 10 -s2: Why people who only give themselves to others. logp(s2|RoBERTa ) =−48.66 logp(s2|GPT-2 ) =−47.13 0 -s1: He healed faster than any professional sports player. logp(s1|ELECTRA ) =−48.77 log p(s1|BERT ) =−50.21 10 -s2: One gets less than a single soccer team. logp(s2|ELECTRA ) =−38.25 logp(s2|BERT ) =−59.09 0 -s1: That is the narrative we have been sold. logp(s1|BERT ) =−56.14 log p(s1|GPT-2 ) =−26.31 10 -s2: This is the week you have been dying. logp(s2|BERT ) =−50.66 logp(s2|GPT-2 ) =−39.50 0 -s1: The resilience is made stronger by early adversity. logp(s1|XLM ) =−62.95 log p(s1|RoBERTa ) =−54.34 10 -s2: Every thing is made alive by infinite Ness. logp(s2|XLM ) =−42.95 logp(s2|RoBERTa ) =−75.72 0 -s1: President Trump threatens to storm the White House. logp(s1|LSTM ) =−58.78 log p(s1|RoBERTa ) =−41.67 10 -s2: West Surrey refused to form the White House. logp(s2|LSTM ) =−40.35 logp(s2|RoBERTa ) =−67.32 0 -s1: Las beans taste best with a mustard sauce. logp(s1|RNN) =−131.62 log p(s1|RoBERTa ) =−60.58 10 -s2: Roughly lanes being alive in a statement ratings. logp(s2|RNN) =−49.31 logp(s2|RoBERTa ) =−99.90 0 -s1: You are constantly seeing people play the multi. logp(s1|3-gram ) =−107.16 log p(s1|ELECTRA ) =−44.79 10 -s2: This will probably the happiest contradicts the hypocrite. logp(s2|3-gram ) =−91.59 logp(s2|ELECTRA ) =−75.83 0 -s1: A buyer can own a genuine product also. logp(s1|2-gram ) =−127.35 log p(s1|ELECTRA ) =−40.21 10 -s2: One versed in circumference of highschool I rambled. logp(s2|2-gram ) =−113.73 logp(s2|ELECTRA ) =−92.61 0 -Table 2: Examples of controversial synthetic-sentence pairs that maximally contributed to each model’s -prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which -the model failed severely. In each case, the failing model 1 prefers sentence s2(higher log probability bolded), -while the model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair -prefer sentence s1. (When more than one sentence pair induced an equal maximal error in a model, the example -included in the table was chosen at random.) -81.0 - 0.5 - 0.0 0.5 1.0 -ordinal correlation between human ratings and models' -sentence pair probability log-ratio (signed-rank cosine similarity)GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -Figure 4: Ordinal correlation of the models’ sentence probability log-ratios and human Likert ratings. -For each sentence pair, model prediction was quantified by logp(s1|m) -p(s2|m). This log-ratio was correlated with the -Likert ratings of each particular participant, using signed-rank cosine similarity (see Methods). This analysis, -taking all trials and human confidence level into account, indicates that GPT-2 performed best in predicting -human sentence probability judgments. However, its predictions are still significantly misaligned with the -human choices. See Fig. 1’s caption for details on the visualization convention. -2.4 Evaluating the entire dataset reveals a hierarchy of models -Rather than evaluating each model’s prediction accuracy with respect to the particular sentence pairs that were -formed to compare this model to alternative models, we can maximize our statistical power by computing the -average prediction accuracy for each model with respect to all of the experimental trials we collected. Fur- -thermore, rather than binarizing the human and model judgments, here we measure the ordinal correspondence -between the graded human choices (taking confidence into account) and the log ratio of the sentence proba- -bilities assigned by each candidate model. Using this more sensitive benchmark (Fig. 4), we found GPT-2 to -be the most human-aligned, followed by RoBERTa; then ELECTRA; BERT; XLM and LSTM; and the RNN, -3-gram, and 2-gram models. However, all of the models (including GPT-2) were found to be significantly less -accurate than the lower bound on the noise ceiling. -One possible reason for the poorer performance of the bidirectional transformers (RoBERTa, ELECTRA, -BERT, and XLM) compared to the unidirectional transformer (GPT-2) is that computing sentence probabilities -in these models is complex, and the probability estimator we developed (see Methods, “Evaluating sentence -probabilities in transformer models”) could be non-optimal; Indeed, the popular pseudo-log-likelihood (PLL) -approach yields slightly higher accuracy for randomly sampled natural-sentence pairs (Extended Data Fig. 5a). -And yet, when we directly compared our estimator to PLL by means of generating and administrating new -synthetic controversial sentences, our estimator was found to be markedly better aligned to human judgments -(Extended Data Fig. 5b and Extended Data Table 2). -Finally, a control analysis employing probability measures normalized by token count revealed that such -normalization had minimal influence on the observed differences among models (Supplementary Section 7.2 -and Supplementary Fig. S1). -93 Discussion -In this study, we probed the ability of language models to predict human relative sentence probability judgments -using controversial sentence pairs, selected or synthesized so that two models disagreed about which sentence -was more probable. We found that (1) GPT-2 (a unidirectional transformer model trained on predicting up- -coming tokens) and RoBERTa (a bidirectional transformer trained on a held-out token prediction task) were the -most predictive of human judgments on controversial natural-sentence pairs (Fig. 1b); (2) GPT-2, RoBERTa, -and ELECTRA (a bidirectional transformer trained on detecting corrupted tokens) were the most predictive of -human judgments on pairs of sentences synthesized to maximize controversiality (Fig. 3a); and (3) GPT-2 was -the most human-consistent model when considering the entire behavioral dataset we collected (Fig. 4). And yet, -all of the models, including GPT-2, exhibited behavior inconsistent with human judgments; using an alternative -model as a counterforce, we could corrupt natural sentences such that their probability under a model did not -decrease, but humans tended to reject the corrupted sentence as unlikely (Fig. 3b). -3.1 Implications for computational psycholinguistic modeling -Unlike convolutional neural networks, whose architectural design principles are roughly inspired by biological -vision [26], the design of current neural network language models is largely uninformed by psycholinguistics -and neuroscience. And yet, there is an ongoing effort to adopt and adapt neural network language models -to serve as computational hypotheses of how humans process language, making use of a variety of different -architectures, training corpora, and training tasks [11, 27–35]. We found that recurrent neural networks make -markedly human-inconsistent predictions once pitted against transformer-based neural networks. This find- -ing coincides with recent evidence that transformers also outperform recurrent networks for predicting neural -responses as measured by ECoG or fMRI [11, 32], as well as with evidence from model-based prediction -of human reading speed [33, 36] and N400 amplitude [36, 37]. Among the transformers, GPT-2, RoBERTa, -and ELECTRA showed the best performance. These models are trained to optimize only word-level pre- -diction tasks, as opposed to BERT and XLM which are additionally trained on next-sentence prediction and -cross-lingual tasks, respectively (and have the same architecture as RoBERTa). This suggests that local word -prediction provides better alignment with human language comprehension. -Despite the agreement between our results and previous work in terms of model ranking, the significant -failure of GPT-2 in predicting the human responses to natural versus synthetic controversial pairs (Fig. 3b) -demonstrates that GPT-2 does not fully emulate the computations employed in human processing of even short -sentences. This outcome is in some ways unsurprising, given that GPT-2 (like all of the other models we con- -sidered) is an off-the-shelf machine learning model that was not designed with human psycholinguistic and -physiological details in mind. And yet, the considerable human inconsistency we observed seems to stand in -stark contrast with the recent report of GPT-2 explaining about 100 percent of the explainable variance in fMRI -and ECoG responses to natural sentences [32]. Part of this discrepancy could be explained by the fact that -Schrimpf and colleagues [32] mapped GPT-2 hidden-layer activations to brain data by means of regularized -linear regression, which can identify a subspace within GPT-2’s language representation that is well-aligned -with brain responses even if GPT-2’s overall sentence probabilities are not human-like. More importantly, -when language models are evaluated with natural language, strong statistical models might capitalize on fea- -tures in the data that are distinct from, but highly correlated with, features that are meaningful to humans. -Therefore, a model that performs well on typical sentences might employ computational mechanisms that are -very distinct from the brain’s, which will only be revealed by testing the model in a more challenging domain. -Note that even the simplest model we considered—a 2-gram frequency table—actually performed quite well -10on predicting human judgments for randomly-sampled natural sentences, and its deficiencies only became ob- -vious when challenged by controversial sentence pairs. We predict that there will be substantial discrepancies -between neural representations and current language models when using stimuli that intentionally stress-test -this relationship, using our proposed sentence-level controversiality approach or complementary ideas such as -maximizing controversial transition probabilities between consecutive words [38]. -Using controversial sentences can be seen as a generalization test of language models: can models predict -what kinds of changes to a natural sentence will lead to humans rejecting the sentence as improbable? Humans -are sometimes capable of comprehending language with atypical constructions (e.g. in cases when pragmatic -judgments can be made about a speaker’s intentions from environmental and linguistic context [39]), but none of -the models we tested were fully able to predict which syntactic or semantic perturbations would be accepted or -rejected by humans. One possibility is that stronger next-word prediction models, using different architectures, -learning rules, or training data, might close the gap between models and humans. Alternatively, it might be -that optimizing for other linguistic tasks, or even non-linguistic task demands (in particular, representing the -external world, the self, and other agents) will turn out to be critical for achieving human-like natural language -processing [40]. -3.2 Controversial sentence pairs as adversarial attacks -Machine vision models are highly susceptible to adversarial examples [41, 42]. Such adversarial examples are -typically generated by choosing a correctly classified natural image and then searching for a minuscule (and -therefore human-imperceptible) image perturbation that would change the targeted model’s classification. The -prospect that similar covert model failure modes may exist also for language models has motivated proposed -generalizations of adversarial methods to textual inputs [43]. However, imperceptible perturbations cannot be -applied to written text: any modified word or character is humanly perceptible. Prior work on adversarial -examples for language models have instead relied on heuristic constraints aiming to limit the change in the -meaning of the text, such as flipping a character [44, 45], changing number or gender [46], or replacing words -with synonyms [47–49]. However, since these heuristics are only rough approximations of human language -processing, many of these methods fail to preserve semantic meaning [50]. Interactive (“human-in-the-loop”) -adversarial approaches allow human subjects to repeatedly alter model inputs such that it confuses target models -but not secondary participants [17, 51], but these approaches are inherently slow and costly and are limited by -mental models the human subjects form about the evaluated language models. -By contrast, testing language models on controversial sentence pairs does not require approximating or -querying a human ground truth during optimization—the objective of controversiality is independent of cor- -rectness. Instead, by designing inputs to elicit conflicting predictions among the models and assessing human -responses to these inputs only once the optimization loop has terminated, we capitalize on the simple fact that -if two models disagree with respect to an input, at least one of the models must be making an incorrect predic- -tion. Pitting language models against other language models also can be conducted by other approaches such -as “red-teaming”, where an alternative language model is used as a generator of potential adversarial examples -for a targeted model and a classifier is used to filter the generated examples such that the output they induce -in the targeted model is indeed incorrect [52]. Our approach shares the underlying principle that an alternative -language model can drive a more powerful test than handcrafted heuristics, but here the models have symmetric -roles (there are no “attacking” and “attacked” models) and we can optimize stimuli directly without relying on -filtering. -113.3 Limitations and future directions -While our results demonstrate that using controversial stimuli can identify subtle differences in language mod- -els’ alignment with human judgments, our study was limited in a number of ways. Our stimuli were all 8-word -English sentences, limiting our ability to make cognitively meaningful claims that apply to language use glob- -ally. 8-word sentences are long enough to include common syntactic constructions and convey meaningful -ideas but may not effectively probe long-distance syntactic dependencies [53]. Future work may introduce -additional sentence lengths and languages, as well as (potentially adaptive) controversial sentence optimization -procedures that consider large sets of candidate models, allowing for greater model coverage than our sim- -pler pairwise approach. Future work may also complement the model-comparative experimental design with -procedures designed to identify potential failure modes common to all models. -A more substantial limitation of the current study is that, like any comparison of pre-trained neural networks -as potential models of human cognition, there could be multiple reasons (i.e., training data, architecture, training -tasks, learning rules) why particular models are better aligned with human judgments. For example, as we -did not systematically control the training corpora used for training the models, it is possible that some of -the observed differences are due to differences in the training sets rather than model architecture. Therefore, -while our results expose failed model predictions, they do not readily answer why these failed predictions arise. -Future experiments could compare custom-trained or systematically manipulated models, which reflect specific -hypotheses about human language processing. In Extended Data Fig. 5, we demonstrate the power of using -synthetic controversial stimuli to conduct sensitive comparisons between models with subtle differences in how -sentence probabilities are calculated. -It is important to note that our analyses considered human relative probability judgments as reflecting -a scalar measure of acceptability. We made this assumption in order to bring the language models (which -assign a probability measure to each sentence) and the human participants onto a common footing. However, -it is possible that different types of sentence pairs engage different human cognitive processes. For pairs of -synthetic sentences, both sentences may be unacceptable in different ways (e.g. exhibit different kinds of -grammatical violations), requiring a judgment that weighs the relative importance of multiple dimensions [54] -and could therefore produce inconsistent rankings across participants or across trials [55]. By contrast, asking -participants to compare a natural and a synthetic sentence (Fig. 3b, Extended Data Table 1) may be more -analogous to previous work measuring human acceptability judgments for sentence pairs [24]. Nonetheless, it -is worth noting that for allof the controversial conditions, the noise ceiling was significantly above the models’ -prediction accuracy, indicating non-random human preferences unexplained by current models that should be -accounted for by future models, which may have to be more complex and capture multiple processes. -Finally, the use of synthetic controversial sentences can be extended beyond probability judgments. A -sufficiently strong language model may enable constraining the experimental design search-space to particular -sentence distributions (e.g., movie reviews or medical questions). Given such a constrained space, we may -be able to search for well-formed sentences that elicit contradictory predictions in alternative domain-specific -models (e.g., sentiment classifiers or question-answering models). However, as indicated by our results, the -task of capturing distributions of well-formed sentences is less trivial than it seems. -124 Methods -4.1 Language models -We tested nine models from three distinct classes: n-gram models, recurrent neural networks, and transformers. -The n-gram models were trained with open source code from the Natural Language Toolkit [56], the recurrent -neural networks were trained with architectures and optimization procedures available in PyTorch [57], and -the transformers were implemented with the open-source repository HuggingFace [58]. For full details see -Supplementary Section 6.1. -4.2 Evaluating sentence probabilities in transformer models -We then sought to compute the probability of arbitrary sentences under each of the models described above. -The term “sentence” is used in this context in its broadest sense—a sequence of English words, not necessarily -restricted to grammatical English sentences. Unlike some classification tasks in which valid model predictions -may be expected only for grammatical sentences (e.g., sentiment analysis), the sentence probability comparison -task is defined over the entire domain of eight-word sequences. -For the set of unidirectional models, evaluating sentence probabilities was performed simply by summing -the log probabilities of each successive token in the sentence from left to right, given all the previous tokens. -For bidirectional models, this process was not as straightforward. One challenge is that transformer model -probabilities do not necessarily reflect a coherent joint probability; the summed log sentence probability result- -ing from adding words in one order (e.g. left to right) does not necessarily equal the probability resulting from a -different order (e.g. right to left). Here we developed a novel formulation of bidirectional sentence probabilities -in which we considered all permutations of serial word positions as possible construction orders (analogous to -the random word visitation order used to sample serial reproduction chains, [59]). In practice, we observed that -the distribution of log probabilities resulting from different permutations tends to center tightly around a mean -value (for example, for RoBERTa evaluated with natural sentences, the average coefficient of variation was -approximately 0.059). Therefore in order to efficiently calculate bidirectional sentence probability, we evaluate -100 different random permutations and define the overall sentence log probability as the mean log probability -calculated from each permutation. Specifically, we initialized an eight-word sentence with all tokens replaced -with the “mask” token used in place of to-be-predicted words during model training. We selected a random -permutation Pof positions 1 through 8, and started by computing the probability of the word at first of these -positions P1given the other seven “mask” tokens. We then replaced the “mask” at position P1with the actual -word at this position and computed the probability of the word at P2given the other six “mask” tokens and the -word at P1. This process was repeated until all “mask” tokens had been filled by the corresponding word. -A secondary challenge in evaluating sentence probabilities in bidirectional transformer models stems from -the fact that these models use word-piece tokenizers (as opposed to whole words), and that these tokenizers are -different for different models. For example, one tokenizer might include the word “beehive” as a single token, -while others strive for a smaller library of unique tokens by evaluating “beehive” as the two tokens “bee” and -“hive”. The model probability of a multi-token word—similar to the probability of a multi-word sentence— -may depend on the order in which the chain rule is applied. Therefore, all unique permutations of token order -for each multi-token word were also evaluated within their respective “masks”. For example, the probability of -the word “beehive” would be evaluated as follows: -13logp(w=beehive ) =0.5 -logp(w1=bee|w2=MASK ) + log p(w2=hive|w1=bee) -+0.5 -logp(w2=hive|w1=MASK ) + log p(w1=bee|w2=hive) (1) -This procedure aimed to yield a more fair estimate of the conditional probabilities of word-piece tokens -and therefore the overall probabilities of multi-token words by 1) ensuring that the word-piece tokens were -evaluated within the same context of surrounding words and masks, and 2) eliminating the bias of evaluating -the word-piece tokens in any one particular order in models which were trained to predict bidirectionally. -One more procedure was applied in order to ensure that all models were computing a probability distribution -over sentences with exactly 8 words. When evaluating the conditional probability of a masked word in models -with word-piece tokenizers, we normalized the model probabilities to ensure that only single words were being -considered, rather than splitting the masked tokens into multiple words. At each evaluation step, each token was -restricted to come from one of four normalized distributions: i) single-mask words were restricted to be tokens -with appended white space, ii) masks at the beginning of a word were restricted to be tokens with preceding -white space (in models with preceding white space, e.g. BERT), iii) masks at the end of words were restricted -to be tokens with trailing white space (in models with trailing white space, e.g. XLM), and iv) masks in the -middle of words were restricted to tokens with no appended white space. -4.3 Assessing potential token count effects on sentence probabilities -Note that, because tokenization schemes varied across models, the number of tokens in a sentence could dif- -fer for different models. These alternative tokenizations can be conceived of as different factorizations of the -modeled language distribution, changing how a sentence’s log probability is additively partitioned across the -conditional probability chain (but not affecting its overall probability) [60]. Had we attempted to normalize -across models by dividing the log probability by the number of tokens, as is often done when aligning model -predictions to human acceptability ratings [12, 13], our probabilities would have become strongly tokenization- -dependent [60]. To empirically confirm that tokenization differences were not driving our results, we sta- -tistically compared the token counts of each model’s preferred synthetic sentences with the token counts of -their non-preferred counterparts. While we found significant differences for some of the models, there was -no systematic association between token count and model sentence preferences (Supplementary Table S1). In -particular, lower sentence probabilities were not systematically confounded by higher token counts. -4.4 Defining a shared vocabulary -To facilitate the sampling, selection, and synthesis of sentences that could be evaluated by all of the candidate -models, we defined a shared vocabulary of 29,157unique words. Defining this vocabulary was necessary in -order to unify the space of possible sentences between the transformer models (which can evaluate any input -due to their word-piece tokenizers) and the neural network and n-gram models (which include whole words as -tokens), and to ensure we only included words that were sufficiently prevalent in the training corpora for all -models. The vocabulary consisted of the words in the subtlex database [61], after removing words that occurred -fewer than 300 times in the 300M word corpus (see Supplementary Section 6.1) used to train the n-gram and -recurrent neural network models (i.e., with frequencies lower than one in a million). -144.5 Sampling of natural sentences -Natural sentences were sampled from the same four text sources used to construct the training corpus for the -n-gram and recurrent neural network models, while ensuring that there was no overlap between training and -testing sentences. Sentences were filtered to include only those with eight distinct words and no punctuation -aside from periods, exclamation points, or question marks at the end of a sentence. Then, all eight-word -sentences were further filtered to include only the words included in the shared vocabulary and to exclude those -included in a predetermined list of inappropriate words and phrases. To identify controversial pairs of natural -sentences, we used integer linear programming to search for sentences that had above-median probability in -one model and minimum probability rank in another model (see Supplementary Section 6.2). -4.6 Generating synthetic controversial sentence pairs -For each pair of models, we synthesized 100 sentence triplets. Each triplet was initialized with a natural -sentence n(sampled from Reddit). The words in sentence nwere iteratively modified to generate a synthetic -sentence with reduced probability according to the first model but not according to the second model. This -process was repeated to generate another synthetic sentence from n, in which the roles of the two models -were reversed. Conceptually, this approach resembles Maximum Differentiation (MAD) competition [62], -introduced to compare models of image quality assessment. Each synthetic sentence was generated as a solution -for a constrained minimization problem: -s∗= argmin -slogp(s|mreject ) -subject to logp(s|maccept )≥logp(n|maccept )(2) -mreject denotes the model targeted to assign reduced sentence probability to the synthetic sentence compared -to the natural sentence, and maccept denotes the model targeted to maintain a synthetic sentence probability -greater or equal to that of the natural sentence. For one synthetic sentence, one model served as maccept and -the other model served as mreject , and for the other synthetic sentence the model roles were flipped. -At each optimization iteration, we selected one of the eight words pseudorandomly (so that all eight posi- -tions would be sampled Ntimes before any position would be sampled N+ 1times) and searched the shared -vocabulary for the replacement word that would minimize the logp(s|mreject )under the constraint. We ex- -cluded potential replacement words that already appeared in the sentence, except for a list of 42 determiners and -prepositions such as “the”, “a”, or “with”, which were allowed to repeat. The sentence optimization procedure -was concluded once eight replacement attempts (i.e., words for which no loss-reducing replacement has been -found) have failed in a row. -4.7 Word-level search for bidirectional models -For models for which the evaluation of logp(s|m)is computationally cheap (2-gram, 3-gram, LSTM, and -the RNN), we directly evaluated the log-probability of the 29,157sentences resulting from each of the 29,157 -possible word replacements. When such probability vectors were available for both models, we simply chose -the replacement minimizing the loss. For GPT-2, whose evaluation is slower, we evaluated sentence probabil- -ities only for word replacements for which the new word had a conditional log-probability (given the previous -words in the sentence) of no less than −10; in rare cases when this threshold yielded fewer than 10 candidate -words, we reduced the threshold in steps of 5until there were at least 10 words above the threshold. For the -15bi-directional models (BERT, RoBERTa, XLM, and ELECTRA), for which the evaluation of logp(s|m)is -costly even for a single sentence, we used a heuristic to prioritize which replacements to evaluate. -Since bi-directional models are trained as masked language models, they readily provide word-level com- -pletion probabilities. These word-level log-probabilities typically have positive but imperfect correlation with -the log-probabilities of the sentences resulting from each potential completion. We hence formed a simple -linear regression-based estimate of logp(s{i} ← w|m), the log-probability of the sentence swith word w -assigned at position i, predicting it from logp(s{i}=w|m, s{i} ←mask ), the completion log-probability -of word wat position i, given the sentence with the i-th word masked: -log ˆp(s{i} ←w|m) =β1logp(s{i}=w|m, s{i} ←mask ) +β0 (3) -This regression model was estimated from scratch for each word-level search. When a word was first -selected for being replaced, the log-probability of two sentences was evaluated: the sentence resulting from -substituting the existing word with the word with the highest completion probability and the sentence resulting -from substituting the existing word with the word with the lowest completion probability. These two word- -sentence log-probability pairs, as well as the word-sentence log-probability pair pertaining to the current word, -were used to fit the regression line. The regression prediction, together with the sentence probability for the -other model (either the exact probability, or approximate probability if the other model was also bi-directional) -was used to predict logp(s|mreject )for each of the 29,157potential replacements. We then evaluated the -true (non-approximate) sentence probabilities of the replacement word with the minimal predicted probability. -If this word indeed reduced the sentence probability, it was chosen to serve as the replacement and the word- -level search was terminated (i.e., proceeding to search a replacement for another word in the sentence). If it -did not reduce the probability, the regression model (Eq. 3) was updated with the new observation, and the -next replacement expected to minimize the sentence probability was evaluated. This word-level search was -terminated after five sentence evaluations that did not reduce the loss. -4.8 Selecting the best triplets from the optimized sentences -Since the discrete hill-climbing procedure described above is highly local, the degree to which this succeeded in -producing highly-controversial pairs varied depending on the starting sentence n. We found that typically, nat- -ural sentences with lower than average log-probability gave rise to synthetic sentences with greater controver- -siality. To better represent the distribution of natural sentences while still choosing the best (most controversial) -triplets for human testing, we used stratified selection. -First, we quantified the controversiality of each triplet as -cm1,m2(n, s1, s2) = logp(n|m1) -p(s1|m1)+ logp(n|m2) -p(s2|m2), (4) -where s1is the sentence generated to reduce the probability in model m1ands2is the sentence generated to -reduce the probability in model m2. -We employed integer programming to choose the 10 most controversial triplets from the 100 triplets opti- -mized for each model pair (maximizing the total controversiality across the selected triplets), while ensuring -that for each model, there was exactly one natural sentence in each decile of the natural sentences probability -distribution. The selected 10 synthetic triplets were then used to form 30 unique experimental trials per model -pair, comparing the natural sentence with one synthetic sentence, comparing the natural sentence with the other -synthetic sentence, and comparing the two synthetic sentences. -164.9 Design of the human experiment -Our experimental procedures were approved by the Columbia University Institutional Review Board (protocol -number IRB-AAAS0252) and were performed in accordance with the approved protocol. All participants -provided informed consent prior. We presented the controversial sentence pairs selected and synthesized by -the language models to 100 native English-speaking, US-based participants (55 male) recruited from Prolific -(www.prolific .co), and paid each participant $5.95. The average participant age was 34.08 ± 12.32. The -subjects were divided into 10 groups, and each ten-subject group was presented with a unique set of stimuli. -Each stimulus set contained exactly one sentence pair from every possible combination of model pairs and -the four main experimental conditions: selected controversial sentence pairs; natural vs. synthetic pair, where -one model served as maccept and the other as mreject ; a natural vs. synthetic pair with the reverse model -role assignments; and directly pairing the two synthetic sentences. These model-pair-condition combinations -accounted for 144 (36 ×4) trials of the task. In addition to these trials, each stimulus set also included nine trials -consisting of sentence pairs randomly sampled from the database of eight-word sentences (and not already -included in any of the other conditions). All subjects also viewed 12 control trials consisting of a randomly -selected natural sentence and the same natural sentence with the words scrambled in a random order. The order -of trials within each stimulus set as well as the left-right screen position of sentences in each sentence pair -were randomized for all participants. While each sentence triplet produced by the optimization procedure (see -subsection “Generating synthetic controversial sentence pairs”) gave rise to three trials, these were allocated -such that no subject viewed the same sentence twice. -On each trial of the task, participants were asked to make a binary decision about which of the two sentences -they considered more probable (for the full set of instructions given to participants, see Supplementary Fig. S2). -In addition, they were asked to indicate one of three levels of confidence in their decision: somewhat confident, -confident, or very confident. The trials were not timed, but a 90-minute time limit was enforced for the whole -experiment. A progress bar at the bottom of the screen indicated to participants how many trials they had -completed and had remaining to complete. -We rejected the data of 21 participants who failed to choose the original, unshuffled sentence in at least -11 of the 12 control trials, and acquired data from 21 alternative participants instead, all of whom passed this -data-quality threshold. In general, we observed high agreement in sentence preferences among our participants, -though the level of agreement varied across conditions. There was complete or near-complete agreement (at -least 9/10 participants with the same binary sentence preference) in 52.2% of trials for randomly-sampled -natural-sentence pairs, 36.6% of trials for controversial natural-sentence pairs, 67.6% of trials for natural- -synthetic pairs, and 60.0% of trials for synthetic-synthetic pairs (versus a chance rate of 1.1%, assuming a -binomial distribution with p= 0.5). -4.10 Evaluation of model-human consistency -To measure the alignment on each trial between model judgments and human judgments, we binarized both -measures; we determined which of the two sentences was assigned with a higher probability by the model, -regardless of the magnitude of the probability difference, and which of the two sentences was favored by the -subject, regardless of the reported confidence level. When both the subject and the model chose the same -sentence, the trial was considered as correctly predicted by that model. This correctness measure was averaged -across sentence pairs and across the 10 participants who viewed the same set of trials. For the lower bound on -the noise ceiling, we predicted each subject’s choices from a majority vote of the nine other subjects who were -presented with the same trials. For the upper bound (i.e., the highest possible accuracy attainable on this data -17sample), we included the subject themselves in this majority vote-based prediction. -Since each of the 10 participant groups viewed a unique trial set, these groups provided 10 independent -replications of the experiment. Models were compared to each other and to the lower bound of the noise -ceiling by a Wilcoxon signed-rank test using these 10 independent accuracy outcomes as paired samples. For -each analysis, the false discovery rate across multiple comparisons was controlled by the Benjamini-Hochberg -procedure [63]. -In Fig. 4, we instead measure model-human consistency in a more continuous way, comparing the sentence -probability ratio in a model to the graded Likert ratings provided by humans; see Supplementary Section 6.3 -for full details. -4.11 Selecting trials for model evaluation -All of the randomly sampled natural-sentence pairs (Fig. 1a) were evaluated for each of the candidate models. -Controversial sentence pairs (either natural, Fig. 1b or synthetic, Fig. 3) were included in a model’s evaluation -set only if they were formed to target that model specifically. The overall summary analysis (Fig. 4) evaluated -all models on all available sentence pairs. -4.12 Comparison to pseudo-log-likelihood acceptability measures -Wang & Cho [64] proposed an alternative approach for computing sentence probabilities in bidirectional -(BERT-like) models, using a pseudo-log-likelihood measure that simply sums the log-probability of each token -conditioned on all of the other tokens in the sentence. While this measure does not reflect a true probability -distribution [65], it is positively correlated with human acceptability judgments for several bidirectional models -[13, 66]. To directly compare this existing approach to our novel method for computing probabilities, we again -used the method of controversial sentence pairs to identify the approach most aligned with human judgments. -For each bidirectional model (BERT, RoBERTa, and ELECTRA), we created two copies of the model, each us- -ing a different approach for computing sentence probabilities. We synthesized 40 sentence pairs to maximally -differentiate between the two copies of each model, with each copy assigning a higher probability to a different -sentence in the pair. Subsequently, we tested 30 human participants, presenting each participant with all 120 -sentence pairs. Model-human consistency was quantified as in the main experiment. -4.13 Data and code availability -The experimental stimuli, detailed behavioral testing results, sentence optimization code, and code for repro- -ducing all analyses and figures are available at github.com/dpmlab/contstimlang [67]. -Acknowledgments -This material is based upon work partially supported by the National Science Foundation under Grant No. -1948004 to NK. This publication was made possible with the support of the Charles H. Revson Foundation to -TG. The statements made and views expressed, however, are solely the responsibility of the authors. -18Author Contributions -T.G., M.S., N.K., and C.B. designed the study. M.S. implemented the computational models and T.G. imple- -mented the sentence pair optimization procedures. M.S. conducted the behavioral experiments. T.G. and M.S. -analyzed the experiments’ results. T.G., M.S., N.K., and C.B. wrote the paper. -Competing Interests -The authors declare no competing interests. -References -1. Rumelhart, D. E., Hinton, G. E. & Williams, R. J. Learning representations by back-propagating errors. -Nature 323, 533–536. doi: 10.1038/323533a0 (1986). -2. Hochreiter, S. & Schmidhuber, J. Long Short-Term Memory. Neural Computation 9, 1735–1780. doi: 10. -1162/neco.1997.9.8.1735 (1997). -3. Devlin, J., Chang, M., Lee, K. & Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers -for Language Understanding inProceedings of the 2019 Conference of the North American Chapter of -the Association for Computational Linguistics: Human Language Technologies (Minneapolis, MN, USA, -2019), 4171–4186. doi: 10.18653/v1/n19-1423 . -4. Liu, Y ., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L. & Stoyanov, -V . RoBERTa: A Robustly Optimized BERT Pretraining Approach. Preprint at https://arxiv.org/ -abs/1907.11692 (2019). -5. Conneau, A. & Lample, G. Cross-lingual Language Model Pretraining inAdvances in Neural Information -Processing Systems 32 (Vancouver, BC, Canada, 2019). URL:proceedings.neurips.cc/paper/ -2019/file/c04c19c2c2474dbf5f7ac4372c5b9af1-Paper.pdf . -6. Clark, K., Luong, M., Le, Q. V . & Manning, C. D. ELECTRA: Pre-training Text Encoders as Discrimina- -tors Rather Than Generators in8th International Conference on Learning Representations, ICLR 2020 -(Online, 2020). URL:openreview.net/forum?id=r1xMH1BtvB . -7. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al. Language models are un- -supervised multitask learners 2019. URL:cdn.openai.com/better- language- models/ -language_models_are_unsupervised_multitask_learners.pdf . -8. Goodkind, A. & Bicknell, K. Predictive power of word surprisal for reading times is a linear function of -language model quality inProceedings of the 8th Workshop on Cognitive Modeling and Computational -Linguistics (CMCL 2018) (Salt Lake City, Utah, 2018), 10–18. doi: 10.18653/v1/W18-0102 . -9. Shain, C., Blank, I. A., van Schijndel, M., Schuler, W. & Fedorenko, E. fMRI reveals language-specific -predictive coding during naturalistic sentence comprehension. Neuropsychologia 138, 107307. doi: 10. -1016/j.neuropsychologia.2019.107307 (2020). -10. Broderick, M. P., Anderson, A. J., Di Liberto, G. M., Crosse, M. J. & Lalor, E. C. Electrophysiologi- -cal correlates of semantic dissimilarity reflect the comprehension of natural, narrative speech. Current -Biology 28, 803–809. doi: 10.1016/j.cub.2018.01.080 (2018). -1911. Goldstein, A., Zada, Z., Buchnik, E., Schain, M., Price, A., Aubrey, B., Nastase, S. A., Feder, A., Emanuel, -D., Cohen, A., Jansen, A., Gazula, H., Choe, G., Rao, A., Kim, C., Casto, C., Fanda, L., Doyle, W., -Friedman, D., Dugan, P., Melloni, L., Reichart, R., Devore, S., Flinker, A., Hasenfratz, L., Levy, O., -Hassidim, A., Brenner, M., Matias, Y ., Norman, K. A., Devinsky, O. & Hasson, U. Shared computational -principles for language processing in humans and deep language models. Nature Neuroscience 25, 369– -380. doi: 10.1038/s41593-022-01026-4 (2022). -12. Lau, J. H., Clark, A. & Lappin, S. Grammaticality, Acceptability, and Probability: A Probabilistic View -of Linguistic Knowledge. Cognitive Science 41, 1202–1241. doi: 10.1111/cogs.12414 (2017). -13. Lau, J. H., Armendariz, C., Lappin, S., Purver, M. & Shu, C. How Furiously Can Colorless Green Ideas -Sleep? Sentence Acceptability in Context. Transactions of the Association for Computational Linguistics -8, 296–310. doi: 10.1162/tacl_a_00315 (2020). -14. Wang, A., Singh, A., Michael, J., Hill, F., Levy, O. & Bowman, S. R. GLUE: A Multi-Task Benchmark -and Analysis Platform for Natural Language Understanding in7th International Conference on Learning -Representations, ICLR 2019, (New Orleans, LA, USA, 2019). URL:openreview.net/forum?id= -rJ4km2R5t7 . -15. Wang, A., Pruksachatkun, Y ., Nangia, N., Singh, A., Michael, J., Hill, F., Levy, O. & Bowman, S. Su- -perGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems inAdvances -in Neural Information Processing Systems 32 (Vancouver, BC, Canada, 2019). URL:proceedings. -neurips.cc/paper/2019/file/4496bf24afe7fab6f046bf4923da8de6-Paper.pdf . -16. Warstadt, A., Parrish, A., Liu, H., Mohananey, A., Peng, W., Wang, S.-F. & Bowman, S. R. BLiMP: The -Benchmark of Linguistic Minimal Pairs for English. Transactions of the Association for Computational -Linguistics 8, 377–392. doi: 10.1162/tacl_a_00321 (2020). -17. Kiela, D., Bartolo, M., Nie, Y ., Kaushik, D., Geiger, A., Wu, Z., Vidgen, B., Prasad, G., Singh, A., Ring- -shia, P., Ma, Z., Thrush, T., Riedel, S., Waseem, Z., Stenetorp, P., Jia, R., Bansal, M., Potts, C. & Williams, -A.Dynabench: Rethinking Benchmarking in NLP inProceedings of the 2021 Conference of the North -American Chapter of the Association for Computational Linguistics: Human Language Technologies -(Online, 2021), 4110–4124. doi: 10.18653/v1/2021.naacl-main.324 . -18. Box, G. E. & Hill, W. J. Discrimination Among Mechanistic Models. Technometrics 9, 57–71. doi: 10. -1080/00401706.1967.10490441 (1967). -19. Golan, T., Raju, P. C. & Kriegeskorte, N. Controversial stimuli: Pitting neural networks against each other -as models of human cognition. Proceedings of the National Academy of Sciences 117, 29330–29337. -doi:10.1073/pnas.1912334117 (2020). -20. Cross, D. V . Sequential dependencies and regression in psychophysical judgments. Perception & Psy- -chophysics 14, 547–552. doi: 10.3758/BF03211196 (1973). -21. Foley, H. J., Cross, D. V . & O’reilly, J. A. Pervasiveness and magnitude of context effects: Evidence -for the relativity of absolute magnitude estimation. Perception & Psychophysics 48, 551–558. doi: 10. -3758/BF03211601 (1990). -22. Petzschner, F. H., Glasauer, S. & Stephan, K. E. A Bayesian perspective on magnitude estimation. Trends -in Cognitive Sciences 19, 285–293. doi: 10.1016/j.tics.2015.03.002 (2015). -23. Greenbaum, S. Contextual Influence on Acceptability Judgments. Linguistics 15. doi: 10.1515/ling. -1977.15.187.5 (1977). -2024. Sch ¨utze, C. T. & Sprouse, J. in (eds Podesva, R. J. & Sharma, D.) 27–50 (Cambridge University Press, -Cambridge, 2014). doi: 10.1017/CBO9781139013734.004 . -25. Sprouse, J. & Almeida, D. Design sensitivity and statistical power in acceptability judgment experiments. -Glossa 2, 14. doi: 10.5334/gjgl.236 (2017). -26. Lindsay, G. W. Convolutional Neural Networks as a Model of the Visual System: Past, Present, and -Future. Journal of Cognitive Neuroscience 33, 2017–2031. doi: 10.1162/jocn_a_01544 (2021). -27. Wehbe, L., Vaswani, A., Knight, K. & Mitchell, T. Aligning context-based statistical models of language -with brain activity during reading inProceedings of the 2014 Conference on Empirical Methods in Nat- -ural Language Processing (EMNLP) (Doha, Qatar, 2014), 233–243. doi: 10.3115/v1/D14-1030 . -28. Toneva, M. & Wehbe, L. Interpreting and improving natural-language processing (in machines) with -natural language-processing (in the brain) inAdvances in Neural Information Processing Systems 32 -(Vancouver, BC, Canada, 2019). URL:proceedings . neurips . cc / paper / 2019 / file / -749a8e6c231831ef7756db230b4359c8-Paper.pdf . -29. Heilbron, M., Armeni, K., Schoffelen, J.-M., Hagoort, P. & De Lange, F. P. A hierarchy of linguistic -predictions during natural language comprehension. Proceedings of the National Academy of Sciences -119, e2201968119. doi: 10.1073/pnas.2201968119 (2022). -30. Jain, S., V o, V ., Mahto, S., LeBel, A., Turek, J. S. & Huth, A. Interpretable multi-timescale models for -predicting fMRI responses to continuous natural speech inAdvances in Neural Information Processing -Systems 33 (Online, 2020), 13738–13749. URL:proceedings.neurips.cc/paper_files/ -paper/2020/file/9e9a30b74c49d07d8150c8c83b1ccf07-Paper.pdf . -31. Lyu, B., Marslen-Wilson, W. D., Fang, Y . & Tyler, L. K. Finding structure in time: Humans, machines, -and language. bioRxiv. Preprint at https://www.biorxiv.org/content/10.1101/2021. -10.25.465687v2 (2021). -32. Schrimpf, M., Blank, I. A., Tuckute, G., Kauf, C., Hosseini, E. A., Kanwisher, N., Tenenbaum, J. B. & -Fedorenko, E. The neural architecture of language: Integrative modeling converges on predictive pro- -cessing. Proceedings of the National Academy of Sciences 118, e2105646118. doi: 10.1073/pnas. -2105646118 (2021). -33. Wilcox, E., Vani, P. & Levy, R. A Targeted Assessment of Incremental Processing in Neural Language -Models and Humans inProceedings of the 59th Annual Meeting of the Association for Computational -Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: -Long Papers) (Online, 2021), 939–952. doi: 10.18653/v1/2021.acl-long.76 . -34. Caucheteux, C. & King, J.-R. Brains and algorithms partially converge in natural language processing. -Communications Biology 5, 134. doi: 10.1038/s42003-022-03036-1 (2022). -35. Arehalli, S., Dillon, B. & Linzen, T. Syntactic Surprisal From Neural Models Predicts, But Underesti- -mates, Human Processing Difficulty From Syntactic Ambiguities inProceedings of the 26th Conference -on Computational Natural Language Learning (CoNLL) (Abu Dhabi, United Arab Emirates (Hybrid), -2022), 301–313. doi: 10.18653/v1/2022.conll-1.20 . -36. Merkx, D. & Frank, S. L. Human Sentence Processing: Recurrence or Attention? Proceedings of the -Workshop on Cognitive Modeling and Computational Linguistics. doi:10.18653/v1/2021.cmcl- -1.2.URL:dx.doi.org/10.18653/v1/2021.cmcl-1.2 (2021). -2137. Michaelov, J. A., Bardolph, M. D., Coulson, S. & Bergen, B. K. Different kinds of cognitive plausibility: -why are transformers better than RNNs at predicting N400 amplitude? inProceedings of the Annual Meet- -ing of the Cognitive Science Society 43 (2021). URL:escholarship.org/uc/item/9z06m20f . -38. Rakocevic, L. I. Synthesizing controversial sentences for testing the brain-predictivity of language models -https://hdl.handle.net/1721.1/130713 . PhD thesis (Massachusetts Institute of Technol- -ogy, 2021). -39. Goodman, N. D. & Frank, M. C. Pragmatic Language Interpretation as Probabilistic Inference. Trends in -Cognitive Sciences 20, 818–829. doi: 10.1016/j.tics.2016.08.005 (2016). -40. Howell, S. R., Jankowicz, D. & Becker, S. A model of grounded language acquisition: Sensorimotor -features improve lexical and grammatical learning. Journal of Memory and Language 53, 258–276. -doi:https://doi.org/10.1016/j.jml.2005.03.002 (2005). -41. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I. & Fergus, R. Intriguing -properties of neural networks Preprint at http://arxiv.org/abs/1312.6199 . 2013. -42. Goodfellow, I. J., Shlens, J. & Szegedy, C. Explaining and Harnessing Adversarial Examples in3rd -International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, -Conference Track Proceedings (2015). URL:arxiv.org/abs/1412.6572 . -43. Zhang, W. E., Sheng, Q. Z., Alhazmi, A. & Li, C. Adversarial attacks on deep-learning models in natural -language processing: A survey. ACM Transactions on Intelligent Systems and Technology (TIST) 11, 1– -41. doi: 10.1145/3374217 (2020). -44. Liang, B., Li, H., Su, M., Bian, P., Li, X. & Shi, W. Deep Text Classification Can be Fooled inProceedings -of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18 (Stockholm, -Sweden, 2018), 4208–4215. doi: 10.24963/ijcai.2018/585 . -45. Ebrahimi, J., Rao, A., Lowd, D. & Dou, D. HotFlip: White-Box Adversarial Examples for Text Classifica- -tioninProceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume -2: Short Papers) (Melbourne, Australia, 2018), 31–36. doi: 10.18653/v1/P18-2006 . -46. Abdou, M., Ravishankar, V ., Barrett, M., Belinkov, Y ., Elliott, D. & Søgaard, A. The Sensitivity of Lan- -guage Models and Humans to Winograd Schema Perturbations inProceedings of the 58th Annual Meet- -ing of the Association for Computational Linguistics (Online, 2020), 7590–7604. doi: 10.18653/v1/ -2020.acl-main.679 . -47. Alzantot, M., Sharma, Y ., Elgohary, A., Ho, B.-J., Srivastava, M. & Chang, K.-W. Generating Natural -Language Adversarial Examples inProceedings of the 2018 Conference on Empirical Methods in Natural -Language Processing (Brussels, Belgium, 2018), 2890–2896. doi: 10.18653/v1/D18-1316 . -48. Ribeiro, M. T., Singh, S. & Guestrin, C. Semantically Equivalent Adversarial Rules for Debugging NLP -models inProceedings of the 56th Annual Meeting of the Association for Computational Linguistics -(Volume 1: Long Papers) (Melbourne, Australia, 2018), 856–865. doi: 10.18653/v1/P18-1079 . -49. Ren, S., Deng, Y ., He, K. & Che, W. Generating Natural Language Adversarial Examples through Prob- -ability Weighted Word Saliency inProceedings of the 57th Annual Meeting of the Association for Com- -putational Linguistics (Florence, Italy, 2019), 1085–1097. doi: 10.18653/v1/P19-1103 . -50. Morris, J., Lifland, E., Lanchantin, J., Ji, Y . & Qi, Y . Reevaluating Adversarial Examples in Natural -Language inFindings of the Association for Computational Linguistics: EMNLP 2020 (Online, 2020), -3829–3839. doi: 10.18653/v1/2020.findings-emnlp.341 . -2251. Wallace, E., Rodriguez, P., Feng, S., Yamada, I. & Boyd-Graber, J. Trick Me If You Can: Human-in-the- -Loop Generation of Adversarial Examples for Question Answering. Transactions of the Association for -Computational Linguistics 7, 387–401. doi: 10.1162/tacl_a_00279 (2019). -52. Perez, E., Huang, S., Song, F., Cai, T., Ring, R., Aslanides, J., Glaese, A., McAleese, N. & Irving, G. -Red Teaming Language Models with Language Models inProceedings of the 2022 Conference on Em- -pirical Methods in Natural Language Processing (Abu Dhabi, United Arab Emirates, 2022), 3419–3448. -doi:10.18653/v1/2022.emnlp-main.225 . -53. Gibson, E. Linguistic complexity: locality of syntactic dependencies. Cognition 68, 1–76. doi: 10.1016/ -S0010-0277(98)00034-1 (1998). -54. Watt, W. C. The indiscreteness with which impenetrables are penetrated. Lingua 37, 95–128. doi: 10. -1016/0024-3841(75)90046-7 (1975). -55. Sch ¨utze, C. T. The empirical base of linguistics. Grammaticality judgments and linguistic methodology -Classics in Linguistics 2.doi:10.17169/langsci.b89.100 (Language Science Press, Berlin, -2016). -56. Bird, S., Klein, E. & Loper, E. Natural language processing with Python: analyzing text with the natural -language toolkit (”O’Reilly Media, Inc.”, Sebastopol, CA, USA, 2009). -57. Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, -N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, -S., Steiner, B., Fang, L., Bai, J. & Chintala, S. PyTorch: An Imperative Style, High-Performance Deep -Learning Library inAdvances in Neural Information Processing Systems 32 (Vancouver, BC, Canada, -2019), 8024–8035. URL:papers.neurips.cc/paper/9015-pytorch-an-imperative- -style-high-performance-deep-learning-library.pdf . -58. Wolf, T., Debut, L., Sanh, V ., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Fun- -towicz, M., Davison, J., Shleifer, S., von Platen, P., Ma, C., Jernite, Y ., Plu, J., Xu, C., Scao, T. L., Gugger, -S., Drame, M., Lhoest, Q. & Rush, A. M. Transformers: State-of-the-Art Natural Language Processing -inProceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System -Demonstrations (Online, 2020), 38–45. doi: 10.18653/v1/2020.emnlp-demos.6 . -59. Yamakoshi, T., Griffiths, T. & Hawkins, R. Probing BERT’s priors with serial reproduction chains in -Findings of the Association for Computational Linguistics: ACL 2022 (Dublin, Ireland, 2022), 3977– -3992. doi: 10.18653/v1/2022.findings-acl.314 . -60. Chestnut, S. Perplexity Accessed: 2022-09-23. 2019. URL:drive.google.com/uc?export= -download&id=1gSNfGQ6LPxlNctMVwUKrQpUA7OLZ83PW . -61. Van Heuven, W. J. B., Mandera, P., Keuleers, E. & Brysbaert, M. Subtlex-UK: A New and Improved Word -Frequency Database for British English. Quarterly Journal of Experimental Psychology 67, 1176–1190. -doi:10.1080/17470218.2013.850521 (2014). -62. Wang, Z. & Simoncelli, E. P. Maximum differentiation (MAD) competition: A methodology for compar- -ing computational models of perceptual quantities. Journal of Vision 8, 8–8. doi: 10.1167/8.12.8 -(2008). -63. Benjamini, Y . & Hochberg, Y . Controlling the False Discovery Rate: A Practical and Powerful Approach -to Multiple Testing. Journal of the Royal Statistical Society: Series B (Methodological) 57, 289–300. -doi:10.1111/j.2517-6161.1995.tb02031.x (1995). -2364. Wang, A. & Cho, K. BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language -Model inProceedings of the Workshop on Methods for Optimizing and Evaluating Neural Language -Generation (Minneapolis, Minnesota, 2019), 30–36. doi: 10.18653/v1/W19-2304 . -65. Cho, K. BERT has a Mouth and must Speak, but it is not an MRF kyunghyuncho.me/bert-has- -a-mouth-and-must-speak-but-it-is-not-an-mrf/ . Accessed: 2022-09-28. 2019. -66. Salazar, J., Liang, D., Nguyen, T. Q. & Kirchhoff, K. Masked Language Model Scoring inProceedings -of the 58th Annual Meeting of the Association for Computational Linguistics (Online, 2020), 2699–2712. -doi:10.18653/v1/2020.acl-main.240 . -67. Golan, T., Siegelman, M., Kriegeskorte, N. & Baldassano, C. Code and data for ”Testing the limits of -natural language models for predicting human language judgments” version 1.2.2. 2023. doi: 10.5281/ -zenodo.8147166 . -68. Shannon, C. E. A mathematical theory of communication. The Bell System Technical Journal 27, 379– -423. doi: 10.1002/j.1538-7305.1948.tb01338.x (1948). -69. Irvine, A., Langfus, J. & Callison-Burch, C. The American Local News Corpus inProceedings of the Ninth -International Conference on Language Resources and Evaluation (LREC’14) (Reykjavik, Iceland, 2014), -1305–1308. URL:www.lrec-conf.org/proceedings/lrec2014/pdf/914_Paper.pdf . -70. Kneser, R. & Ney, H. Improved backing-off for m-gram language modeling in1995 international confer- -ence on acoustics, speech, and signal processing 1 (1995), 181–184. doi: 10.1109/ICASSP.1995. -479394 . -71. Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual 2021. URL:www.gurobi.com . -72. Woodbury, M. A. Rank Correlation when There are Equal Variates. The Annals of Mathematical Statistics -11, 358–362. URL:www.jstor.org/stable/2235684 (1940). -73. Sch ¨utt, H. H., Kipnis, A. D., Diedrichsen, J. & Kriegeskorte, N. Statistical inference on representational -geometries. eLife 12 (eds Serences, J. T. & Behrens, T. E.) e82566. doi: 10.7554/eLife.82566 -(2023). -74. Nili, H., Wingfield, C., Walther, A., Su, L., Marslen-Wilson, W. & Kriegeskorte, N. A Toolbox for Rep- -resentational Similarity Analysis. PLOS Computational Biology 10, 1–11. doi: 10.1371/journal. -pcbi.1003553 (2014). -75. Pennington, J., Socher, R. & Manning, C. GloVe: Global Vectors for Word Representation inProceedings -of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Doha, Qatar, -2014), 1532–1543. doi: 10.3115/v1/D14-1162 . -76. Frank, S. L. & Willems, R. M. Word predictability and semantic similarity show distinct patterns of -brain activity during language comprehension. Language, Cognition and Neuroscience 32, 1192–1203. -doi:10.1080/23273798.2017.1323109 (2017). -5 Extended Data -24Extended Data Figure 1: An example of one experimental trial, as presented to the participants . The participant must -choose one sentence while providing their confidence rating on a 3-point scale. -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 2 -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 10.93 0.88 0.86 0.83 0.87 0.76 0.82 0.78 -0.93 0.90 0.83 0.88 0.89 0.80 0.84 0.82 -0.88 0.90 0.87 0.84 0.90 0.83 0.90 0.86 -0.86 0.83 0.87 0.84 0.83 0.77 0.83 0.77 -0.83 0.88 0.84 0.84 0.86 0.81 0.79 0.77 -0.87 0.89 0.90 0.83 0.86 0.87 0.91 0.84 -0.76 0.80 0.83 0.77 0.81 0.87 0.84 0.89 -0.82 0.84 0.90 0.83 0.79 0.91 0.84 0.91 -0.78 0.82 0.86 0.77 0.77 0.84 0.89 0.91 -0.00 0.25 0.50 0.75 1.00 -between model agreement rate -(proportion of sentence pairs) -Extended Data Figure 2: Between-model agreement rate -on the probability ranking of the 90 randomly sampled -and paired natural sentence pairs evaluated in the exper- -iment . Each cell represents the proportion of sentence pairs -for which two models make congruent probability ranking -(i.e., both models assign a higher probability to sentence 1, -or both models assign a higher probability to sentence 2). -25GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 2 -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 10.51 0.53 0.78 0.83 0.66 0.84 0.79 0.88 -0.49 0.60 0.65 0.78 0.60 0.80 0.76 0.86 -0.47 0.40 0.63 0.75 0.57 0.79 0.64 0.59 -0.22 0.35 0.37 0.58 0.30 0.44 0.60 0.54 -0.17 0.22 0.25 0.42 0.47 0.49 0.42 0.55 -0.34 0.40 0.43 0.70 0.53 0.84 0.62 0.70 -0.16 0.20 0.21 0.56 0.51 0.16 0.49 0.62 -0.21 0.24 0.36 0.40 0.58 0.38 0.51 0.63 -0.12 0.14 0.41 0.46 0.45 0.30 0.38 0.37 -0.00 0.25 0.50 0.75 1.00 -human choice aligned with model 1 -(proportion of trials) -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 2 -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodel 10.55 0.61 0.79 0.84 0.85 0.86 0.76 0.92 -0.45 0.36 0.60 0.89 0.81 0.96 0.94 0.95 -0.39 0.64 0.49 0.88 0.89 0.95 0.98 0.95 -0.21 0.40 0.51 0.71 0.73 0.86 0.78 0.90 -0.16 0.11 0.12 0.29 0.51 0.70 0.72 0.91 -0.15 0.19 0.11 0.27 0.49 0.79 0.65 0.71 -0.14 0.04 0.05 0.14 0.30 0.21 0.32 0.55 -0.24 0.06 0.02 0.22 0.28 0.35 0.68 0.67 -0.08 0.05 0.05 0.10 0.09 0.29 0.45 0.33 -0.00 0.25 0.50 0.75 1.00 -human choice aligned with model 1 -(proportion of trials)aNatural controversial sentences bSynthetic controversial sentencesExtended Data Figure 3: Pairwise model comparison of model-human consistency. For each pair of models (represented -as one cell in the matrices above), the only trials considered were those in which the stimuli were either selected (a) or -synthesized (b) to contrast the predictions of the two models. For these trials, the two models always made controversial -predictions (i.e., one sentence is preferred by the first model and the other sentence is preferred by the second model). The -matrices above depict the proportion of trials in which the binarized human judgments aligned with the row model (“model -1”). For example, GPT-2 (top-row) was always more aligned (green hues) with the human choices than its rival models. In -contrast, 2-gram (bottom-row) was always less aligned (purple hues) with the human choices than its rival models. -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodels assigned as mreject -GPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-grammodels assigned as maccept0.21 0.24 0.43 0.23 0.29 0.36 0.17 0.22 -0.21 0.16 0.29 0.51 0.29 0.23 0.18 0.39 -0.17 0.20 0.27 0.34 0.27 0.40 0.27 0.20 -0.14 0.18 0.27 0.44 0.23 0.25 0.33 0.20 -0.07 0.07 0.11 0.16 0.17 0.13 0.08 0.22 -0.03 0.01 0.07 0.04 0.17 0.06 0.16 0.04 -0.05 0.03 0.05 0.03 0.03 0.01 0.04 0.06 -0.04 0.03 0.02 0.08 0.05 0.04 0.06 0.04 -0.01 0.02 0.01 0.02 0.04 0.01 0.07 0.04 -0.00 0.25 0.50 0.75 1.00 -humans choice aligned with maccept -(proportion of trials) -Extended Data Figure 4: Pairwise model analysis of human re- -sponse for natural vs. synthetic sentence pairs. In each optimiza- -tion condition, a synthetic sentence swas formed by modifying a nat- -ural sentence nso the synthetic sentence would be “rejected” by one -model ( mreject , columns), minimizing p(s|mreject ), and would be -“accepted” by another model ( maccept , rows), satisfying the constraint -p(s|maccept )≥p(n|maccept ). Each cell above summarizes model- -human agreement in trials resulting from one such optimization condi- -tion. The color of each cell denotes the proportion of trials in which -humans judged a synthetic sentence to be more likely than its natural -counterpart and hence aligned with maccept . For example, the top-right -cell depicts human judgments for sentence pairs formed to minimize -the probability assigned to the synthetic sentence by the simple 2-gram -model while ensuring that GPT-2 would judge the synthetic sentence -to be at least as likely as the natural sentence; humans favored the syn- -thetic sentence in only 22 out the 100 sentence pairs in this condition. -261.0 -0.5 -0.0 0.5 1.0 -ordinal correlation between human ratings and models' -sentence pair probability log-ratio (signed-rank cosine similarity)RoBERTa -RoBERTa (PLL) -ELECTRA -ELECTRA (PLL) -BERT -BERT (PLL) -aRandomly sampled natural-sentence pairs -1.0 -0.5 -0.0 0.5 1.0 -ordinal correlation between human ratings and models' -sentence pair probability log-ratio (signed-rank cosine similarity)RoBERTa -RoBERTa (PLL) -ELECTRA -ELECTRA (PLL) -BERT -BERT (PLL) -bSynthetic controversial sentence pairsExtended Data Figure 5: Human consistency of bidirectional transformers: approximate log-likelihood versus pseudo- -log-likelihood (PLL). Each dot in the plots above depicts the ordinal correlation between the judgments of one participant -and the predictions of one model. (a)The performance of BERT, RoBERTa, and ELECTRA in predicting the human judg- -ments of randomly sampled natural sentence pairs in the main experiment, using two different likelihood measures: our novel -approximate likelihood method (i.e., averaging multiple conditional probability chains, see Methods) and pseudo-likelihood -(PLL, summating the probability of each word given all of the other words [64]). For each model, we statistically com- -pared the two likelihood measures to each other and to the noise ceiling using a two-sided Wilcoxon signed-rank test across -the participants. False discovery rate was controlled at q <0.05for the 9 comparisons. When predicting human pref- -erences of natural sentences, the pseudo-log-likelihood measure is at least as accurate as our proposed approximate -log-likelihood measure. (b) Results from a follow-up experiment, in which we synthesized synthetic sentence pairs for each -of the model pairs, pitting the two alternative likelihood measures against each other. Statistical testing was conducted in the -same fashion as in panel a. These results indicate that for each of the three bidirectional language models, the approximate -log-likelihood measure is considerably and significantly ( q <0.05) more human-consistent than the pseudo-likelihood mea- -sure. Synthetic controversial sentence pairs uncover a dramatic failure mode of the pseudo-log-likelihood measure, -which remains covert when the evaluation is limited to randomly-sampled natural sentences. See Extended Data Table -2 for synthetic sentence pair examples. -270 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -Extended Data Figure 6: Model prediction accuracy for pairs of -natural and synthetic sentences, evaluating each model across -all of the sentence pairs in which it was targeted to rate the syn- -thetic sentence to be less probable than the natural sentence. -The data binning applied here is complementary to the one used in -Fig. 3b, where each model was evaluated across all of the sentence -pairs in which it was targeted to rate the synthetic sentence to be -at least as probable as the natural sentence. Unlike Fig. 3b, where -all of the models performed poorly, here no models were found to -be significantly below the lower bound on the noise ceiling; typ- -ically, when a sentence was optimized to decrease its probability -under any model (despite the sentence probability not decreasing -under a second model), humans agreed that the sentence became -less probable. -sentence log probability (model 1) log probability (model 2) # human choices -n: I always cover for him and make excuses. logp(n|GPT-2 ) =−36.46 log p(n|2-gram ) =−106.95 10 -s: We either wish for it or ourselves do. logp(s|GPT-2 ) =−36.15 logp(s|2-gram ) =−122.28 0 -n: This is why I will never understand boys. logp(n|RoBERTa ) =−46.88 log p(n|2-gram ) =−103.11 10 -s: This is why I will never kiss boys. logp(s|RoBERTa ) =−46.75 logp(s|2-gram ) =−107.91 0 -n: One of the ones I did required it. logp(n|ELECTRA ) =−35.97 log p(n|LSTM ) =−40.89 10 -s: Many of the years I did done so. logp(s|ELECTRA ) =−35.77 logp(s|LSTM ) =−46.25 0 -n: There were no guns in the Bronze Age. logp(n|BERT ) =−48.48 log p(n|ELECTRA ) =−30.40 10 -s: There is rich finds from the Bronze Age. logp(s|BERT ) =−48.46 logp(s|ELECTRA ) =−44.34 0 -n: You did a great job on cleaning them. logp(n|XLM ) =−40.38 log p(n|RNN) =−43.47 10 -s: She did a great job at do me. logp(s|XLM ) =−39.89 logp(s|RNN) =−61.03 0 -n: This logic has always seemed flawed to me. logp(n|LSTM ) =−39.77 log p(n|RNN) =−45.92 10 -s: His cell has always seemed instinctively to me. logp(s|LSTM ) =−38.89 logp(s|RNN) =−62.81 0 -s: Stand near the cafe and sip your coffee. logp(s|RNN) =−65.55 log p(s|ELECTRA ) =−34.46 10 -n: Sit at the front and break your neck. logp(n|RNN) =−44.18 logp(n|ELECTRA ) =−34.65 0 -n: Most of my jobs have been like this. logp(n|3-gram ) =−80.72 log p(n|LSTM ) =−35.07 10 -s: One of my boyfriend have been like this. logp(s|3-gram ) =−80.63 logp(s|LSTM ) =−41.44 0 -n: They even mentioned that I offer white flowers. logp(n|2-gram ) =−113.38 log p(n|BERT ) =−62.81 10 -s: But even fancied that would logically contradictory philosophies. logp(s|2-gram ) =−113.24 logp(s|BERT ) =−117.98 0 -Extended Data Table 1: Examples of pairs of synthetic and natural sentences that maximally contributed to each -model’s prediction error. For each model (double row, “model 1”), the table shows results for two sentences on which the -model failed severely. In each case, the failing model 1 prefers synthetic sentence s(higher log probability bolded), while the -model it was pitted against (“model 2”) and all 10 human subjects presented with that sentence pair prefer natural sentence -n. (When more than one sentence pair induced an equal maximal error in a model, the example included in the table was -chosen at random.) -28sentence pseudo-log-likelihood (PLL) approximate log probability # human choices -s1: I found so many in things and called. logp(s1|BERT (PLL) ) =−55.14 log p(s1|BERT ) =−55.89 30 -s2: Khrushchev schizophrenic so far -disproportionately goldfish fished alone. logp(s2|BERT (PLL) ) =−22.84 logp(s2|BERT ) =−162.31 0 -s1: Figures out if you are on the lead. logp(s1|BERT (PLL) ) =−38.11 log p(s1|BERT ) =−51.27 30 -s2: Neighbours unsatisfactory indistinguishable -misinterpreting schizophrenic on homecoming -cheerleading. logp(s2|BERT (PLL) ) =−16.43 logp(s2|BERT ) =−258.91 0 -s1: I just say this and not the point. logp(s1|ELECTRA (PLL) ) =−34.41 log p(s1|ELECTRA ) =−33.80 30 -s2: Glastonbury reliably mobilize disenfranchised -homosexuals underestimate unhealthy skeptics. logp(s2|ELECTRA (PLL) ) =−11.81 logp(s2|ELECTRA ) =−162.62 0 -s1: And diplomacy is more people to the place. logp(s1|ELECTRA (PLL) ) =−62.81 log p(s1|ELECTRA ) =−47.33 30 -s2: Brezhnev ingenuity disembarking Acapulco -methamphetamine arthropods unaccompanied -Khrushchev. logp(s2|ELECTRA (PLL) ) =−34.00 logp(s2|ELECTRA ) =−230.97 0 -s1: Sometimes what looks and feels real to you. logp(s1|RoBERTa (PLL) ) =−36.58 log p(s1|RoBERTa ) =−51.61 30 -s2: Buying something breathes or crawls -aesthetically to decorate. logp(s2|RoBERTa (PLL) ) =−9.78 logp(s2|RoBERTa ) =−110.27 0 -s1: In most other high priority packages were affected. logp(s1|RoBERTa (PLL) ) =−71.13 log p(s1|RoBERTa ) =−61.60 30 -s2: Stravinsky cupboard nanny contented burglar -babysitting unsupervised bathtub. logp(s2|RoBERTa (PLL) ) =−21.86 logp(s2|RoBERTa ) =−164.70 0 -Extended Data Table 2: Examples of controversial synthetic-sentence pairs that maximally contributed to the pre- -diction error of bidirectional transformers using pseudo-log-likelihood (PLL). For each bidirectional model, the table -displays two sentence pairs on which the model failed severely when its prediction was based on pseudo-log-likelihood -(PLL) estimates [64]. In each of these sentence pairs, the PLL estimate favors sentence s2(higher PLL bolded), while the -approximate log-likelihood estimate and most of the human subjects presented with that sentence pair preferred sentence -s1. (When more than one sentence pair induced an equal maximal error in a model, the example included in the table was -chosen at random.) Sentences with long, multi-token words (e.g., “methamphetamine”) have high PLL estimates since -each of their tokens is well predicted by the others tokens. And yet, the entire sentence is improbable according to -human judgments and approximate log-probability estimates based on proper conditional probability chains. -296 Supplementary Methods -6.1 Language models -N-gram models. N-gram models [68], the simplest language model class, are trained by counting the -number of occurrences of all unique phrases of length N words in large text corpora. N-gram models make -predictions about upcoming words by using empirical conditional probabilities in the training corpus. We -tested both 2-gram and 3-gram variants. In 2-gram models, all unique two-word phrases are counted, and -each upcoming word probability (probability of w2conditioned on previous word w1) is determined by -dividing the count of 2-gram w1, w2by the count of unigram (word) w1. In 3-gram models, all unique three- -word phrases are counted, and upcoming word probabilities (probability of w3conditioned on previous -words w1andw2) are determined by dividing the count of 3-gram w1, w2, w3by the count of 2-gram -w1, w2. In both such models, sentence probabilities can be computed as the product of all unidirectional -word transition probabilities in a given sentence. We trained both the 2-gram and 3-gram models on a large -corpus composed of text from four sources: 1. public comments from the social media website Reddit -(reddit .com) acquired using the public API at pushshift .io, 2. articles from Wikipedia, 3. English -books and poetry available for free at Project Gutenberg ( gutenberg .org), and 4. articles compiled in -the American Local News Corpus [69]. The n-gram probability estimates were regularized by means of -Kneser-Ney smoothing [70]. -Recurrent neural network models. We also tested two recurrent neural network models, including a -simple recurrent neural network (RNN) [1] and a more complex long short-term memory recurrent neural -network (LSTM) [2]. We trained both of these models on a next word prediction task using the same corpus -used to train the n-gram models. Both the RNN and LSTM had a 256-feature embedding size and a 512- -feature hidden state size, and were trained over 100 independent batches of text for 50 epochs with a learning -rate of .002. Both models’ training sets were tokenized into individual words and consisted of a vocabulary -of94,607unique tokens. -Transformer models. Similar to RNNs, transformers are designed to make predictions about sequential -inputs. However, transformers do not use a recurrent architecture, and have a number of more complex -architectural features. For example, unlike the fixed token embeddings in classic RNNs, transformers utilize -context-dependent embeddings that vary depending on a token’s position. Most transformers also contain -multiple attention heads in each layer of the model, which can help direct the model to relevant tokens in -complex ways. We tested five models with varying architectures and training procedures, including BERT -[3], RoBERTa [4], XLM [5], ELECTRA [6], and GPT-2 [7]. -• We used the large version of BERT (bi-directonal encoder representations from transformers), con- -taining 24 encoding layers, 1024 hidden units in the feedforward network element of the model, and -16 attention heads. BERT is a bi-directional model trained to perform two different tasks: 1. a masked -language modeling (MLM) task, in which 15 percent of tokens are replaced with a special [MASK] -token and BERT must predict the masked word, and 2. next sentence prediction (NSP), in which -BERT aims to predict the upcoming sentence in the training corpus given the current sentence. -• RoBERTa is also a bi-directional model that uses the same architecture as BERT. However, RoBERTa -was trained on exclusively the masked word prediction task (and not next sentence prediction), and -used a different optimization procedure (including longer training on a larger dataset). This makes -empirical comparisons between BERT and RoBERTa particularly interesting, because they differ only -in training procedure and not architecture. -30• XLM is a cross-lingual bi-directional model which, too, shares BERT’s original architecture. XLM is -trained on three different tasks: 1. the same MLM task used in both BERT and RoBERTa, 2. a causal -language modeling task where upcoming words are predicted from left to right, and 3. a translation -modeling task. On this task, each training example consists of the same text in two languages, and the -model performs a masked language modeling task using context from one language to predict tokens -of another. Such a task can help the XLM model become robust to idiosyncrasies of one particular -language that may not convey much linguistic information. -• The ELECTRA model uses a training approach that involves two transformer models: a generator -and a discriminator. While the generator performs a masked language modeling task similar to other -transformers, the discriminator simultaneously tries to figure out which masked tokens were replaced -by the generator. This task may be more efficient than pure masked token prediction, because it uses -information from all input tokens rather than only the masked subset. -• GPT-2, the second iteration of GPT OpenAI’s GPT model, is the only unidirectional transformer -model that we tested. We used the pretrained GPT-2-xl version, with 48 encoding layers and 25 -attention heads in each layer. Because GPT-2 is unidirectional it was trained only on the causal -language modeling task, in which tokens are predicted from left to right. -6.2 Selection of controversial natural-sentence pairs -We evaluated 231,725eight-word sentences sampled from Reddit. Reddit comments were scraped from -across the entire website and all unique eight-word sentences were saved. These sentences were subse- -quently filtered to exclude blatant spelling errors, inappropriate language, and individual words that were -not included in the corpus used to train the n-gram and recurrent neural network models in our experiment. -We estimated logp(s|m)for each natural sentence sand each model mas described above. We -then rank-transformed the sentence probabilities separately for each model, assigning the fractional rank -r(s|m) = 0 to the least probable sentence according to model mandr(s|m) = 1 to the most probable -one. This step eliminated differences between models in terms of probability calibration. -Next, we aimed to filter this corpus for controversial sentences. To prune the candidate sentences, we -eliminated any sentence sfor which no pair of models m1,m2held(r(s|m1)<0.5)and(r(s|m2)≥0.5), -where r(s|m1)is the fractional rank assigned for sentence sby model m. This step ensured that all of the -remaining sentences had a below-median probability according to one model and above-median probability -according to another, for at least one pair of models. We also excluded sentences in which any word (except -for prepositions) appeared more than once. After this pruning, 85,749candidate sentences remained, from -which85749 -2 -≈3.67×109possible sentence pairs can be formed. -We aimed to select 360 controversial sentence pairs, devoting 10 sentence pairs to each of the 36 -models pairs. First, we defined two 360-long integer vectors m1andm2, specifying for each of the -360 yet unselected sentence pairs which model pair they contrast. We then selected 360 sentence pairs -s1 -1, s2 -1 -, -s1 -2, s2 -2 -..., -s1 -360, s2 -360 -by solving the following minimization problem: -{(s1 -j∗, s2 -j∗)|j= 1,2, ..360}= argmin -s1,s2X -j -r(s1 -j|m1 -j) +r(s2 -j|m2 -j) -(5) -subject to ∀jr(s1 -j|m2 -j)≥0.5 (5a) -∀jr(s2 -j|m1 -j)≥0.5 (5b) -All 720 sentences are unique. (5c) -31To achieve this, we used integer linear programming (ILP) as implemented by Gurobi [71]. We repre- -sented sentence allocation as a sparse binary tensor Sof dimensions 85,749 ×360×2 (sentences, trials, -pair members) and the fractional sentence probabilities ranks as a matrix Rof dimensions 85,749 ×9 (sen- -tences, models). This enabled us to express and solve the selection problem in Eq. 5 as a standard ILP -problem: -S∗= argmin -SX -i,jSi,j,1Ri,m1 -j+Si,j,2Ri,m2 -j(6) -subject to Si,j,1Ri,m2 -j≥0.5 (6a) -Si,j,2Ri,m1 -j≥0.5 (6b) -∀iX -j,kSi,j,k≤1(each sentence iis used only once in the experiment) (6c) -∀jP -iSi,j,1= 1 -∀jP -iSi,j,2= 1) -(each trial jis allocated exactly one sentence pair) (6d) -Sis binary (6e) -6.3 Evaluation of model-human consistency: Correlating model log-probability -ratios to human Likert ratings -For every model mand experimental trial i, we evaluated the log probability ratio for the trial’s two sen- -tences: -LR(s1 -i, s2 -i|m) = logp(s2 -i|m) -p(s1 -i|m)(7) -The human Likert ratings were recoded to be symmetrical around zero, mapping the six ratings appear- -ing in Extended Data Fig. 1 to (−2.5,−1.5,−0.5,+0.5,+1.5,+2.5). We then sought to correlate the model -log-ratios and with the zero-centered human Likert ratings, quantifying how well the model log-ratios were -associated with human sentence-likeliness judgments. To allow for an ordinal (not necessarily linear) asso- -ciation between the log-ratios and Likert ratings, we rank-transformed both measures (ranking within each -model or each human) while retaining the sign of the values. -For each participant h: -r(s1 -i, s2 -i|h) = sign( y0(s1 -i, s2 -i|h))·R y0(s1 -i, s2 -i|h)  -, (8) -where y0(s1 -i, s2 -i|h))is the zero-centered Likert rating provided by subject hfor trial iandR(·)is rank -transform using random tie-breaking. -For each model m: -r(s1 -i, s2 -i|m) = sign( LR(s1 -i, s2 -i|m))·R LR(s1 -i, s2 -i|m)  -, (9) -A valid correlation measure of the model ranks and human ranks must be invariant to whether one -sentence was presented on the left ( s1) and the other on the right ( s2), or vice versa. Changing the sentence -order within a trial would flip the signs of both the log-ratio and the zero-centered Likert rating. Therefore, -the required correlation measure must be invariant to such coordinated sign flips, but not to flipping the sign -of just one of the measures. Since cosine similarity maintains such invariance, we introduced signed-rank -32cosine similarity , an ordinal analog of cosine similarity, substituting the raw data points for signed ranks (as -defined in Eq. 8-9): -SCSR=P -ir(s1 -i, s2 -i|m)r(s1 -i, s2 -i|h)qP -ir(s1 -i, s2 -i|m)2qP -ir(s1 -i, s2 -i|h)2. (10) -To eliminate the noise contributed by random tie-breaking, we used a closed-form expression of the -expected value of Eq. 10 over different random tie-breaking draws: -E(SCSR) =P -iE -r(s1 -i, s2 -i|m) -E -r(s1 -i, s2 -i|h) -pPn -k=1k2pPn -k=1k2=P -i¯r(s1 -i, s2 -i|m)¯r(s1 -i, s2 -i|h)Pn -k=1k2, (11) -where ¯r(·)denotes signed rank with average-rank assigned to ties instead of random tie-breaking, and n -denotes the number of evaluated sentence pairs. The expected value of the product in the numerator is -equal to the product of expected values of the factors since the random tie-breaking within each factor is -independent. The vector norms (the factors in the denominator) are constant since given no zero ratings, -each signed-rank rating vector always includes one of each rank 1ton(where nis the number of sentence -pairs considered), and the signs are eliminated by squaring. This derivation follows a classical result for -Spearman’s ρ[72] (see [73], appendix C, for a modern treatment). We empirically confirmed that averaging -SCSR as defined in Eq. 10 across a large number of random tie-breaking draws converges to E(SCSR)as -defined in Eq. 11. This latter expression (whose computation requires no actual random tie-breaking) was -used to quantify the correlation between each participant and model. -For each participant, the lower bound on the noise ceiling was calculated by replacing the model-derived -predictions with an across-participants average of the nine other participants’ signed-rank rating vectors. The -lower bound plotted in main text Fig. 4 is an across-subject average of this estimate. An upper bound on the -noise ceiling was calculated as a dot product between the participant’s expected signed-rank rating vector -(¯ r/pPk2) and a normalized, across-participants average of the expected signed-rank rating vectors of all -10 participants. -Inference was conducted in the same fashion as that employed for the binarized judgments (Wilcoxon -signed-rank tests across the 10 subject groups, controlling for false discovery rate). -7 Supplementary Results -7.1 Randomly sampled natural-sentence pairs fail to adjudicate among mod- -els -As a baseline, we created 90 pairs of natural sentence pairs by randomly sampling from a corpus of 8-word -sentences appearing on Reddit. Evaluating the sentence probabilities assigned to the sentences by the dif- -ferent models, we found that models tended to agree on which of the two sentences was more probable -(Extended Data Fig. 2). The between-model agreement rate ranged from 75.6% of the sentence pairs for -GPT-2 vs. RNN to 93.3% for GPT-2 vs. RoBERTa, with an average agreement between models of 84.5%. -Main text Fig. 1a (left-hand panel) provides a detailed graphical depiction of the relationship between sen- -tence probability ranks for one model pair (GPT-2 and RoBERTa). -We divided these 90 pairs into 10 sets of nine sentences and presented each set to a separate group of 10 -subjects. To evaluate model-human alignment, we computed the proportion of trials where the model and the -participant agreed on which sentence was more probable. All of the nine language models performed above -33chance (50% accuracy) in predicting the human choices for the randomly sampled natural sentence pairs -(main text Fig. 1a, right-hand panel). Since we presented each group of 10 participants with a unique set of -sentence pairs, we could statistically test between-model differences while accounting for both participants -and sentence pairs as random factors by means of a simple two-sided Wilcoxon signed-rank test conducted -across the 10 participant groups. For the set of randomly sampled natural-sentence pairs, this test yielded -no significant prediction accuracy differences between the candidate models (controlling for false discovery -rate for all 36 model pairs at q<.05). This result is unsurprising considering the high level of between- -model agreement on the sentence probability ranking within each of these sentence pairs. -To obtain an estimate of the noise ceiling [74] (i.e., the best possible prediction accuracy for this dataset), -we predicted each participant’s choices by the majority vote of the nine other participants who were pre- -sented with the same trials. This measurement provided a lower bound on the noise ceiling. Including the -participant’s own choice in the prediction yields an upper bound, since no set of predictions can be more -human-aligned on average given the between-subject variability. For the randomly sampled natural sen- -tences, none of the models were found to be significantly less accurate than the lower bound on the noise -ceiling (controlling the false discovery rate for all nine models at q<.05). In other words, the 900 trials -of randomly sampled and paired natural sentences provided no statistical evidence that any of the language -models are human-inconsistent. -7.2 Comparing the accuracy of unnormalized and normalized sentence prob- -ability estimates -Previous studies (e.g., [13]) have found that normalizing language model sentence probability estimates by -the sentences’ token counts can result in greater alignment with human acceptability judgments. While we -deliberately used unnormalized sentence log probability when designing the experiment, we evaluated the -prediction accuracy of each model under such normalizations through a control analysis. -Rather than predicting human judgments based only on the relative log probabilities of the two sen- -tences, we instead used cross-validated logistic regression to predict human judgments using a combination -of unnormalized log probability differences (“LP”) and two measures from Lau and colleagues [13] that -incorporate information about the token counts in each sentence. The “MeanLP” measure normalized each -sentence’s log probability by its token count T, whereas the “PenLP” measure divided each sentence’s log -probability by a dampened version of its token count, -(5 +T)/(5 + 1)0.8. For models trained on whole -words (LSTM, RNN, 3-gram, and 2-gram), we used character count instead of token count. -For each language model m, we fitted a separate logistic regression to predict the individual binarized -sentence choices across the entire main experiment dataset by weighing the three predictors “LP,” “MeanLP,” -and “PenLP.” We did not include an intercept due to the symmetry of the prediction task (the presentation -of sentences as sentence 1 or 2 was randomized). We cross-validated the logistic regression’s accuracy by -leaving out one sentence pair at a time, using data from all conditions of the experiment. -Taking token count into consideration led to minor improvements in the prediction accuracy of most -models (an average improvement of 0.95%), but this adjustment did not change the hierarchy of the models -in terms of their human consistency (Supplementary Fig. S1). We hypothesize that the greater disparities -between unnormalized and normalized probability measures, as observed by Lau and colleagues [13] com- -pared to those found in our study, may be attributed to their experiment involving sentences of markedly -different lengths. -347.3 Models differ in their sensitivity to low-level linguistic features -While the controversial sentences presented in this study were synthesized without consideration for partic- -ular linguistic features, we performed a post hoc analysis to explore the contribution of different features -to model and human preferences (Supplementary Fig. S3). For each controversial synthetic sentence pair, -we computed the average log-transformed word frequency for each sentence (extracted from the publicly -available subtlex database [61]). We also computed the average pairwise correlation between semantic -GloVe vector representations [75] of all eight words, based on neuroimaging research showing that there -are specific neural signatures evoked by dissimilarity in semantic vectors [10, 76]. We performed paired -sample t-tests across sentence pairs between the linguistic feature preferences for models vs. humans, and -found that GPT-2, LSTM, RNN, 3-gram, and 2-gram models were significantly more likely (vs. humans) -to prefer sentences with low GloVe correlations, while ELECTRA was significantly more likely to prefer -high GloVe correlations (controlling the false discovery rate for all nine models at q<.05). For word fre- -quency, the RNN, 3-gram, and 2-gram models were significantly biased (vs. humans) to prefer sentences -with low-frequency words, while ELECTRA and XLM showed a significant bias for high-frequency words. -These results indicate that even strong models like GPT-2 and ELECTRA can exhibit subtle misalignments -with humans in their response to simple linguistic features, when evaluated on sentences synthesized to be -controversial. -358 Supplementary Figures -0 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -a -0 25% 50% 75% 100% -human-choice prediction accuracyGPT-2 -RoBERTa -ELECTRA -BERT -XLM -LSTM -RNN -3-gram -2-gram -bUnnormalized sentence probability estimates -Normalized sentence probability estimates -Supplementary Fig. S1: The predictivity of normalized and unnormalized log-probability measures. -(a)Predicting human judgments from all conditions using only unnormalized log probability differences -(equivalent to Fig. 4 in the main text, except using binarized accuracy as a dependent measure). (b)Binarized -accuracy of the logistic regression optimally combining LP, MeanLP, and PenLP for each language model. -Relative model performance is nearly identical in these two analyses, indicating that tokenization differences -across models did not play a large confounding role in our main results. -36Supplementary Fig. S2: The task instructions provided to the participants at the beginning of the experimen- -tal session. -37abSupplementary Fig. S3: Linguistic feature values for synthetic sentence pairs. (a) GloVe correlation -values of the preferred and rejected sentence for each synthetic sentence pair. Each panel depicts preferences -for both humans (red) and a specific model (black), for sentence pairs that this model was involved in -synthesizing. Black sub-panel outlines indicate significant differences between the preferences of models -and humans on that particular set of sentence pairs, according to a paired sample t-test (controlling for false -discovery rate across all nine models at q<.05). (b)Same as (a), but for average log-transformed word -frequency. -389 Supplementary Tables -modelaccepted sentence -has more tokensequal -token-countsrejected sentence -has more tokens p-value -GPT-2 24 13 3 <0.0001 -RoBERTa 6 18 16 0.0656 -ELECTRA 12 21 7 0.3593 -BERT 4 8 28 <0.0001 -XLM 2 16 22 <0.0001 -Supplementary Table S1: Token count control analysis. For each transformer model, we considered syn- -thetic controversial sentence pairs where the other targeted model was also a transformer (a total of 40 -sentence pairs per model). For each such pair, we evaluated the token count of the synthetic sentence to -which the model assigned a higher probability (“accepted sentence”) and the token count of the synthetic -sentence to which the model assigned a lower probability (“rejected sentence”). For each model, this table -presents the number of sentence pairs in which the accepted sentence had a higher token count, both sen- -tences had an equal number of tokens, and the rejected sentence had a higher token count. We compared the -prevalence of higher token counts in accepted and rejected sentences using a binomial test ( H0:p= 0.5) -controlled for False Discovery Rate across five comparisons. -GPT-2 assigned significantly more tokens to accepted sentences, whereas BERT and XLM assigned sig- -nificantly more tokens to rejected sentences. For RoBeRTa and ELECTRA, no significant difference was -found. Note that since the controversial sentences are driven by relative model response properties, a sig- -nificant difference for a particular model does not necessarily indicate that token count biases the model’s -sentence probability estimates. For example, GPT-2’s apparent preference for sentences with a greater token -count might reflect biases of the alternative models pitted against GPT-2. These models might prefer shorter -sentences that exhibit undetected grammatical or semantic violations over longer but felicitous sentences. -Overall, these results indicate that while certain models’ probability estimates might be biased by to- -kenization, lower sentence probabilities were not systematically confounded by higher token counts. -39 \ No newline at end of file