{"id":1211,"date":"2022-08-31T12:59:25","date_gmt":"2022-08-31T12:59:25","guid":{"rendered":"https:\/\/resources.illc.uva.nl\/illc-blog\/?p=1211"},"modified":"2022-09-05T09:46:07","modified_gmt":"2022-09-05T09:46:07","slug":"probing-by-priming-what-do-large-language-models-know-about-grammar","status":"publish","type":"post","link":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/","title":{"rendered":"Probing by Priming: What do Large Language Models know about Grammar?"},"content":{"rendered":"\n<p class=\"has-small-font-size\">19 July 2022, <a rel=\"noreferrer noopener\" href=\"https:\/\/jumelet.ai\/\" target=\"_blank\">Jaap Jumelet<\/a> <em>with help from co-authors <a href=\"https:\/\/j-anie.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Arabella Sinclair<\/a>, <a href=\"https:\/\/staff.fnwi.uva.nl\/w.zuidema\/\" target=\"_blank\" rel=\"noreferrer noopener\">Willem Zuidema<\/a>, and <a href=\"https:\/\/staff.fnwi.uva.nl\/r.fernandezrovira\/\" target=\"_blank\" rel=\"noreferrer noopener\">Raquel Fern\u00e1ndez<\/a><\/em><\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><a href=\"https:\/\/i.imgur.com\/TNEb1os.png\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"856\" src=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post-1024x856.png\" alt=\"\" class=\"wp-image-1227\" srcset=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post-1024x856.png 1024w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post-300x251.png 300w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post-768x642.png 768w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post-1200x1003.png 1200w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-post.png 1424w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure><\/div>\n\n\n\n<p class=\"has-text-align-right has-small-font-size\"><em>Look closely! This Escherian representation directly reflects our priming paradigm!<\/em><br>\u00a9 <a href=\"http:\/\/mdhk.net\/\">Marianne de Heer Kloots<\/a><\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"alignright size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"983\" height=\"703\" src=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/Screenshot-from-2022-08-31-16-17-49-1.png\" alt=\"\" class=\"wp-image-1237\" srcset=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/Screenshot-from-2022-08-31-16-17-49-1.png 983w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/Screenshot-from-2022-08-31-16-17-49-1-300x215.png 300w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/Screenshot-from-2022-08-31-16-17-49-1-768x549.png 768w\" sizes=\"auto, (max-width: 983px) 100vw, 983px\" \/><figcaption>You can check out the full paper on arXiv here: <a href=\"https:\/\/arxiv.org\/pdf\/2109.14989.pdf\">https:\/\/arxiv.org\/pdf\/2109.14989.pdf<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<p class=\"has-normal-font-size\">With the rise of extremely large language models such as GPT-3<sup>[1]<\/sup> and PaLM<sup>[2]<\/sup>, and their impressive fluency in generating language, the field of natural language processing (NLP) is increasingly asking which rules of grammar and abstract linguistic classes they have really learned, and, what the right way of finding out the answers to this question even are. In our work, recently published in the <em>Transactions of the Association for Computational Linguistics<\/em>, we take inspiration from the field of psycholinguistics. This field has studied the same questions about humans, found out how hard it can be to probe for linguistic concepts directly, and developed a variety of useful tools. In particular, \u2018priming\u2019 is a key tool in psycholinguistics that turns out to be very useful for our purposes as well.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Priming<\/h2>\n\n\n\n<p class=\"has-normal-font-size\">Priming, and in particular <strong>structural priming, <\/strong>has played a central role in demonstrating experimentally that humans possess abstract knowledge of the structure of sentences. The systematic procedure to study structural priming was introduced by J. Kathryn Bock in her seminal 1986 paper \u201c<em>Syntactic Persistence in Language Production<\/em>\u201d<sup>[3]<\/sup>. Structural priming is a phenomenon in which humans tend to unintentionally repeat syntactic constructions that they have been exposed to earlier. For example, if you let someone read the following sentence (called the <em>prime<\/em>):<\/p>\n\n\n\n<pre class=\"wp-block-verse\"><em>A rock star sold an undercover agent some cocaine.<\/em><\/pre>\n\n\n\n<p class=\"has-normal-font-size\">And next, request that person to describe the following image (called the <em>target<\/em>):<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"224\" height=\"335\" src=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-06-30-16-49-24.png\" alt=\"\" class=\"wp-image-1214\" srcset=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-06-30-16-49-24.png 224w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-06-30-16-49-24-201x300.png 201w\" sizes=\"auto, (max-width: 224px) 100vw, 224px\" \/><\/figure><\/div>\n\n\n\n<p>Then it turns out that they are significantly more likely to describe the target image as \u201c<em>A man is reading a boy a book<\/em>\u201d (a <em>double-object dative<\/em>), as opposed to \u201c<em>A man is reading a book to a boy<\/em>\u201d (a <em>prepositional dative<\/em>). The reverse would have been true if the initial sentence that we showed would have been \u201c<em>A rock star sold some cocaine to an undercover agent<\/em>\u201d. We call this phenomenon <strong>structural priming<\/strong>: reading the first sentence <em>primed<\/em> our behaviour on the target task on a structural level.Structural priming is therefore taken as evidence that abstract information, such as the double-object dative in our example, plays an active role in human language processing. Many subsequent studies have examined the various factors that impact this priming behaviour. For example, priming effects increase if the prime and target sentences are more similar to each other, or when the exposure to prime sentences is increased.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Language Models<\/h2>\n\n\n\n<p>Language models are trained to predict the next word based on an input prompt. Nowadays these models are large neural networks (often based on the successful Transformer architecture), and trained on billions of sentences. Their success has led to a dramatic paradigm shift within NLP: it has now become standard practice to first <em>pre-train<\/em> a language model on a large corpus of text, before <em>fine-tuning<\/em> the model on a more specific task of interest. Language models have become incredibly powerful in recent years, up to the point that they may even fool people into believing, for a while, that they are interacting with a real human being.<\/p>\n\n\n\n<p>In our research, we are interested in approaching language models from a similar point of view as the field of psycholinguistics approaches human language processing. We examine the question of <em>how<\/em> language models reason about language, and whether they exhibit similar behavioural patterns as humans do. Since we are interested in finding out whether language models encode abstract structural information in their representations, it is a natural step to apply the procedure of structural priming to these models.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Measuring Priming Behaviour<\/h2>\n\n\n\n<p>In order to measure structural priming behaviour in language models, we need to make some modifications to our experimental setup. Recall that in the original experiments by Bock (1986) humans were asked to describe an image, after having been exposed to a prime sentence. The language models that we are interested in cannot process images, and can hence only be exposed to text. We instead measure a model\u2019s likelihood of a target sentence directly, <em>conditioned<\/em> on a prime sentence of a certain structural form. We then compare a model\u2019s likelihood of a target sentence when conditioned on a prime of the same structural form (e.g. <em>active<\/em>) versus a prime of the opposite form (e.g. <em>passive<\/em>). This leads to the following definition of the <strong>Priming Effect<\/strong>:<\/p>\n\n\n\n<pre class=\"wp-block-verse\"><em>PE<\/em> = <sub>log<\/sub><em>P<\/em>(target<em><sub>X<\/sub><\/em> | prime<em><sub>X<\/sub><\/em>) \u2014 <sub>log<\/sub>P(target<em><sub>X<\/sub><\/em> | prime<em><sub>Y<\/sub><\/em>)<\/pre>\n\n\n\n<p>Let\u2019s look at a concrete example to see how this is operationalised. For the structural forms&nbsp;<em>X<\/em>&nbsp;and&nbsp;<em>Y<\/em>&nbsp;we take the&nbsp;<em>active<\/em>&nbsp;and&nbsp;<em>passive voice<\/em>, and target<sub><em>X<\/em><\/sub>&nbsp;will be&nbsp;<em>\u201cThe man threw the ball<\/em>\u201d. The active prime sentence prime<em><sub>X<\/sub><\/em>&nbsp;will be&nbsp;<em>\u201cThe dog ate the food\u201d<\/em>, and its passive counterpart then becomes \u201c<em>The food was eaten by the dog\u201d<\/em>. To compute the model\u2019s likelihoods, we feed the model one type of prime sentence, and conditioned on that sentence we compute the likelihood of the target. The likelihood of the target is computed by feeding the target sentence one word at a time, and computing the probability of the next word. We do the same steps for the other type of prime sentence, and after subtracting the two likelihood scores we obtain the Priming Effect.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Creating an Evaluation Corpus<\/h2>\n\n\n\n<p>To evaluate the priming behaviour of a range of language models we create a corpus of template-based prime\/target pairs. A major confound that we need to control for when creating this corpus is the overlap between the prime and target sentence. We must avoid that the target is being primed solely because of lexical overlap: if the model only assigns a higher likelihood to the target because the words in the target occurred in the same configuration in the prime we can\u2019t speak of structural priming yet. For structural priming to occur the model must exhibit this behaviour on a more abstract level, i.e. only due to <em>structural overlap<\/em> the model prefers the target conditioned on the congruent prime sentence.<\/p>\n\n\n\n<p>We therefore construct an evaluation corpus in which we systematically control for the relationship between the prime and target sentences. We create multiple settings where prime and target are completely different except for their structure, and additional versions where we vary the <em>lexical, semantic<\/em>, and <em>structural<\/em> similarity. Next to that, we create versions which contain varying numbers of prime sentences to measure whether priming is <em>cumulative <\/em>(more exposure to a prime results in stronger priming)<em>,<\/em> and versions which vary the distance between prime and target to measure the effect of <em>recency <\/em>(more recent exposure to a prime results in stronger priming). All combined, this yields a suite of corpora of around 1.3 million prime\/target pairs that we call <strong>PrimeLM <\/strong>(available <a href=\"https:\/\/github.com\/dmg-illc\/prime-lm\" target=\"_blank\" rel=\"noreferrer noopener\">here<\/a>).<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Language Models Exhibit Structural Priming<\/h2>\n\n\n\n<p>Our experimental setup is divided into two parts: we have a <em>core<\/em> corpus that we use to measure the general priming behaviour, and specialised corpora that are used to measure various factors related to priming like semantic similarity between prime and target and the cumulativity &amp; recency effects. We evaluate the presence of structural priming using our corpus on a range of language models, focusing in particular on GPT-2.&nbsp;<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"773\" src=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57-1024x773.png\" alt=\"\" class=\"wp-image-1217\" srcset=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57-1024x773.png 1024w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57-300x227.png 300w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57-768x580.png 768w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57-1200x906.png 1200w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/Screenshot-from-2022-07-19-16-57-57.png 1286w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n\n<p><strong>We find evidence that models are susceptible to structural priming, in a manner similar to humans. <\/strong>As an example, we show in the figure above the results of the recency and cumulativity experiments. The left plot shows that the closer a Prime (P<sub>X<\/sub>) is to the target (from left to right, right being closest), the stronger the Priming Effect becomes. We increase the distance of the prime and the target by inserting intransitive sentences (P<sub>Z<\/sub>) in between. For example, the <strong>P<sub>X<\/sub><\/strong> P<sub>Z<\/sub> P<sub>Z<\/sub> P<sub>Z<\/sub> template would translate to a sequence such as:<\/p>\n\n\n\n<pre class=\"wp-block-verse\"><em>A king kills a mayor. He can exist. They did go. You will exist.&nbsp;<\/em><\/pre>\n\n\n\n<p>Followed by a target sentence: <em>The nurse forgot the chicken. <\/em>The priming effect turns out to be the strongest when the prime is closest to the target sentence (P<sub>Z<\/sub> P<sub>Z<\/sub> P<sub>Z <\/sub><strong>P<\/strong><strong><sub>X<\/sub><\/strong>).<\/p>\n\n\n\n<p>The right plot shows that the Priming Effect increases proportionally to the number of prime sentences that are fed to the model, before computing the target likelihood. This means that the language models assign a higher likelihood to a target sentence if they have been exposed to several examples of sentences with the same structure (the more the better), even if these earlier sentences have no words or even meanings in common with the target sentence! This provides evidence that the models are truly prone to <em>structural <\/em>priming: the only information that is repeated between prime and target occurs on a structural level.<\/p>\n\n\n\n<p>Both these effects correspond to human behaviour in priming experiments with recency and cumulativity manipulations.<sup>[4][5]<\/sup> Other experiments, related to semantic factors of structural priming and structural complexity, can be found in the paper along with our analysis.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Outlook<\/h2>\n\n\n\n<p>Understanding precisely what language models have learnt, and how this influences the language they produce can help us both in understanding how they will perform when used in downstream applications, e.g. identifying possible sources of errors, and in creating more human-like models of language.&nbsp;<\/p>\n\n\n\n<p>One of the main approaches to understanding intelligence is through behavioural studies, e.g. in a given context with a given stimulus, how does the human or AI react? Paradigms from psycholinguistics, which&nbsp; study language processing in humans, can be applied to evaluating the language abilities of language models, testing to what extent they display certain linguistic behaviour in certain contexts, such as the evidence of structural priming in our work. As such, we hope that our study paves a way for a more comprehensive integration of psycholinguistics and NLP.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"alignright size-medium\"><img loading=\"lazy\" decoding=\"async\" width=\"300\" height=\"251\" src=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/dieuwke_hupkes-1024x856-1-300x251.png\" alt=\"\" class=\"wp-image-1219\" srcset=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/dieuwke_hupkes-1024x856-1-300x251.png 300w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/dieuwke_hupkes-1024x856-1-768x642.png 768w, https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/07\/dieuwke_hupkes-1024x856-1.png 1024w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><figcaption><em>Lightening up the black box<\/em>: Read in <a rel=\"noreferrer noopener\" href=\"https:\/\/resources.illc.uva.nl\/illc-blog\/lightening-up-the-black-box\/\" target=\"_blank\">this earlier blogpost<\/a> about related work on interpretability within the ILLC.<\/figcaption><\/figure><\/div>\n\n\n\n<p>Machine learning models are often referred to as <a rel=\"noreferrer noopener\" href=\"https:\/\/resources.illc.uva.nl\/illc-blog\/lightening-up-the-black-box\/\" target=\"_blank\">black boxes<\/a>. The fields of Interpretability and Explainable AI seek to allow us to \u2018look under the hood\u2019 and try to explore the reasoning behind an AI model\u2019s decisions. This can be thought of as parallel to what neuroscientists do when they examine patterns of brain activity and try to relate it to the behaviour of an animal, or to what molecular biologists do when investigating how a particular medicine actually interacts with the biochemistry of an animal cell. More experiments in this area can help to shed light on what linguistic properties are indeed learnt via the language modelling task. This in turn can lead to identifying areas for improvement, and help us understand what is driving the behaviour of these models.<\/p>\n\n\n\n<hr class=\"wp-block-separator\"\/>\n\n\n\n<p>Interested in reading more about this study? You can check out the paper in more detail on <a rel=\"noreferrer noopener\" href=\"https:\/\/arxiv.org\/pdf\/2109.14989.pdf\" target=\"_blank\">arXiv<\/a>. Do you want to dive directly into the code to see how we implemented our setup in practice? All our scripts and data can be found in <a rel=\"noreferrer noopener\" href=\"https:\/\/github.com\/dmg-illc\/prime-lm\" target=\"_blank\">this repository<\/a>. <\/p>\n\n\n\n<h6 class=\"wp-block-heading\">References<\/h6>\n\n\n\n<p class=\"has-small-font-size\"><sup>[1]<\/sup> Brown, Tom, et al. &#8220;Language models are few-shot learners.&#8221; <em>Advances in neural information processing systems<\/em> 33 (2020): 1877-1901.<br><sup>[2] <\/sup>Chowdhery, Aakanksha, et al. &#8220;Palm: Scaling language modeling with pathways.&#8221; <em>arXiv preprint<\/em> <em>arXiv:2204.02311<\/em> (2022).<br><sup>[3] <\/sup>Bock, J. Kathryn. &#8220;Syntactic persistence in language production.&#8221; <em>Cognitive Psychology<\/em> 18 (1986): 355-387.<br><sup>[4]<\/sup> Kaschak, Michael P. et al. \u201cStructural priming as implicit learning: Cumulative priming effects and individual differences.\u201d <em>Psychonomic Bulletin &amp; Review <\/em>18(6) (2011): 1133\u20131139.<br><sup>[5]<\/sup> Reitter, David et al. \u201cA computational cognitive model of syntactic priming.\u201d <em>Cognitive Science<\/em> 35(4) (2011): 587\u2013637.<\/p>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The rise of large language models raises questions about what rules of grammar they have really learned. Do they have a similar understanding of language as humans? Priming may provide answers.<\/p>\n","protected":false},"author":2,"featured_media":1228,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[19,34,15],"tags":[],"class_list":["post-1211","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-artificial-intelligence","category-linguistics","category-natural-language-processing"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v25.9 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog\" \/>\n<meta property=\"og:description\" content=\"The rise of large language models raises questions about what rules of grammar they have really learned. Do they have a similar understanding of language as humans? Priming may provide answers.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\" \/>\n<meta property=\"og:site_name\" content=\"ILLC Blog\" \/>\n<meta property=\"article:published_time\" content=\"2022-08-31T12:59:25+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2022-09-05T09:46:07+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1024-socials-preview.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1024\" \/>\n\t<meta property=\"og:image:height\" content=\"536\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:image\" content=\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1024-socials-preview.png\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"10 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\"},\"author\":{\"name\":\"admin\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/e6a636d976565eaec71567ead9a5e70a\"},\"headline\":\"Probing by Priming: What do Large Language Models know about Grammar?\",\"datePublished\":\"2022-08-31T12:59:25+00:00\",\"dateModified\":\"2022-09-05T09:46:07+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\"},\"wordCount\":1898,\"publisher\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization\"},\"image\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png\",\"articleSection\":[\"Artificial Intelligence\",\"Linguistics\",\"Natural Language Processing\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\",\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\",\"name\":\"Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog\",\"isPartOf\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png\",\"datePublished\":\"2022-08-31T12:59:25+00:00\",\"dateModified\":\"2022-09-05T09:46:07+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage\",\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png\",\"contentUrl\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png\",\"width\":1424,\"height\":1190},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Probing by Priming: What do Large Language Models know about Grammar?\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#website\",\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/\",\"name\":\"ILLC Blog\",\"description\":\"Institute for Logic, Language and Computation\",\"publisher\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization\",\"name\":\"ILLC Blog\",\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2020\/04\/logo-uva.png\",\"contentUrl\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2020\/04\/logo-uva.png\",\"width\":301,\"height\":30,\"caption\":\"ILLC Blog\"},\"image\":{\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/logo\/image\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/e6a636d976565eaec71567ead9a5e70a\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/9da35c76f4cff342883f387bb36fd693eeeceb9d4c368ae306f833f4b54db32e?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/9da35c76f4cff342883f387bb36fd693eeeceb9d4c368ae306f833f4b54db32e?s=96&d=mm&r=g\",\"caption\":\"admin\"},\"url\":\"https:\/\/resources.illc.uva.nl\/illc-blog\/author\/admin\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/","og_locale":"en_US","og_type":"article","og_title":"Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog","og_description":"The rise of large language models raises questions about what rules of grammar they have really learned. Do they have a similar understanding of language as humans? Priming may provide answers.","og_url":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/","og_site_name":"ILLC Blog","article_published_time":"2022-08-31T12:59:25+00:00","article_modified_time":"2022-09-05T09:46:07+00:00","og_image":[{"width":1024,"height":536,"url":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1024-socials-preview.png","type":"image\/png"}],"author":"admin","twitter_card":"summary_large_image","twitter_image":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1024-socials-preview.png","twitter_misc":{"Written by":"admin","Est. reading time":"10 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#article","isPartOf":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/"},"author":{"name":"admin","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/e6a636d976565eaec71567ead9a5e70a"},"headline":"Probing by Priming: What do Large Language Models know about Grammar?","datePublished":"2022-08-31T12:59:25+00:00","dateModified":"2022-09-05T09:46:07+00:00","mainEntityOfPage":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/"},"wordCount":1898,"publisher":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization"},"image":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage"},"thumbnailUrl":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png","articleSection":["Artificial Intelligence","Linguistics","Natural Language Processing"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/","url":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/","name":"Probing by Priming: What do Large Language Models know about Grammar? - ILLC Blog","isPartOf":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage"},"image":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage"},"thumbnailUrl":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png","datePublished":"2022-08-31T12:59:25+00:00","dateModified":"2022-09-05T09:46:07+00:00","breadcrumb":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#primaryimage","url":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png","contentUrl":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2022\/08\/structuralpriming-1424-thumbnail.png","width":1424,"height":1190},{"@type":"BreadcrumbList","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/probing-by-priming-what-do-large-language-models-know-about-grammar\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/resources.illc.uva.nl\/illc-blog\/"},{"@type":"ListItem","position":2,"name":"Probing by Priming: What do Large Language Models know about Grammar?"}]},{"@type":"WebSite","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#website","url":"https:\/\/resources.illc.uva.nl\/illc-blog\/","name":"ILLC Blog","description":"Institute for Logic, Language and Computation","publisher":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/resources.illc.uva.nl\/illc-blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#organization","name":"ILLC Blog","url":"https:\/\/resources.illc.uva.nl\/illc-blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/logo\/image\/","url":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2020\/04\/logo-uva.png","contentUrl":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-content\/uploads\/2020\/04\/logo-uva.png","width":301,"height":30,"caption":"ILLC Blog"},"image":{"@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/e6a636d976565eaec71567ead9a5e70a","name":"admin","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/resources.illc.uva.nl\/illc-blog\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/9da35c76f4cff342883f387bb36fd693eeeceb9d4c368ae306f833f4b54db32e?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/9da35c76f4cff342883f387bb36fd693eeeceb9d4c368ae306f833f4b54db32e?s=96&d=mm&r=g","caption":"admin"},"url":"https:\/\/resources.illc.uva.nl\/illc-blog\/author\/admin\/"}]}},"_links":{"self":[{"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/posts\/1211","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/comments?post=1211"}],"version-history":[{"count":15,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/posts\/1211\/revisions"}],"predecessor-version":[{"id":1242,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/posts\/1211\/revisions\/1242"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/media\/1228"}],"wp:attachment":[{"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/media?parent=1211"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/categories?post=1211"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/resources.illc.uva.nl\/illc-blog\/wp-json\/wp\/v2\/tags?post=1211"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}