Goldberg, A. E. & Ferreira, F. (2022). Good-enough language production. Trends in Cognitive Science.
Our ability to comprehend and produce language is one of humans’ most impressive skills, but it is not flawless. We must convey and interpret messages via a noisy channel in ever-changing contexts and we sometimes fail to access an optimal combination of words and grammatical constructions. Here, we extend the notion of good-enough (GN) comprehension to GN production, which allows us to unify a wide range of phenomena including overly vague word choices, agreement errors, resumptive pronouns, transfer effects, and children’s overextensions and regularizations. We suggest these all involve the accessing and production of a ‘GN’ option when a more-optimal option is inaccessible. The role of accessibility highlights the need to relate memory encoding and retrieval processes to language comprehension and production.
Mon, S. K., Nencheva, M., Citron, F. M. M., Lew-Williams, C., & Goldberg, A. E. (2021). Conventional metaphors elicit greater real-time engagement than literal paraphrases or concrete sentences. Journal of Memory and Language.
Conventional metaphors (e.g., a firm grasp on an idea) are extremely common. A possible explanation for their ubiquity is that they are more engaging, evoking more focused attention, than their literal paraphrases (e.g., a good understanding of an idea). To evaluate whether, when, and why this may be true, we created a new database of 180 English sentences consisting of conventional metaphors, literal paraphrases, and concrete descriptions (e.g., a firm grip on a doorknob). Extensive norming matched differences across sentence types in complexity, plausibility, emotional valence, intensity, and familiarity of the key phrases. Then, using pupillometry to study the time course of metaphor processing, we predicted that metaphors would elicit greater event-evoked pupil dilation compared to other sentence types. Results confirmed the predicted increase beginning at the onset of the key phrase and lasting seconds beyond the end of the sentence. When metaphorical and literal sentences were compared directly in survey data, participants judged metaphorical sentences to convey “richer meaning,” but not more information. We conclude that conventional metaphors are more engaging than literal paraphrases or concrete sentences in a way that is irreducible to difficulty or ease, amount of information, short-term lexical access, or downstream inferences.
Goldberg, A.E. & Lee, C. (2021). Accessibility and historical change: An emergent cluster led uncles and aunts to become aunts and uncles. Frontiers.
There are times when a curiously odd relic of language presents us with a thread, which when pulled, reveals deep and general facts about human language. This paper unspools such a case. Prior to 1930, English speakers uniformly preferred male-before-female word order in conjoined nouns such as uncles and aunts; nephews and nieces; men and women. Since then, at least a half dozen items have systematically reversed their preferred order (e.g., aunts and uncles, nieces and nephews) while others have not (men and women). We review evidence that the unusual reversals began with mother and dad(dy) and spread to semantically and morphologically related binomials over a period of decades. The present work proposes that three aspects of cognitive accessibility combine to quantify the probability of A&B order: 1) the relative accessibility of the A and B terms individually, 2) competition from B&A order, and critically, 3) cluster strength (i.e., similarity to related A’& B’ cases). The emergent cluster of female-first binomials highlights the influence of semantic neighborhoods in memory retrieval. We suggest that cognitive accessibility can be used to predict the word order of both familiar and novel binomials generally, as well as the diachronic change focused on here.
Goldberg, A. E. & Herbst, T. (2021). The Nice-of-you construction and its fragments. Linguistics.
This article argues that a usage-based construction (a conventional pairing of form and function) is required to account for a certain pattern of English exemplified by e.g., it’s nice of you to read this. Contemporary corpus and survey data reveal that the construction is strongly associated with certain adjectives (e.g., nice, good) over others, while diachronic data demonstrate that the construction’s overall frequency has systematically waxed and waned over the past century. The construction’s unique function – namely to concisely convey a judgment regarding how an action reflects on the agent of the action – enables us to predict many observations about its distribution without stipulation. These include restrictions on the interpretation of adjectives that occur in the construction, its infinitive complement, the modal verbs that may appear in it and its ability to be embedded. We further observe that certain conventional fragments of the construction evoke the semantics of the entire construction. Finally, we situate the construction within a network of related constructions, as part of a speaker’s construct-i-con.
Floyd, S., Jeppsen, C., & Goldberg, A. E. (2020). Children on the Autism spectrum are challenged by complex word meanings. Journal of Autism and Developmental Disorders.
The current work suggests that two factors conspire to make vocabulary learning challenging for youth on the Autism spectrum: (1) a tendency to focus on specifics rather than on relationships among entities, and (2) up to 80% of words are associated with distinct but related meanings (e.g. a baseball cap, pen cap, bottle cap). Neurotypical (NT) children find it easier to learn multiple related meanings of words (polysemy) in comparison to multiple unrelated meanings (homonymy). We exposed 60 NT children and 40 verbal youth on the Autism spectrum to novel words. The groups’ performance learning homonyms was comparable, but unlike their NT peers, youth on the spectrum did not display the same advantage for learning polysemous words compared to homonyms.
Tachihara, K. & Goldberg, A. E. (2020). Cognitive Accessibility predicts word order of couples’ names in English and Japanese. Cognitive Linguistics.
Citron, F. M. M., Michaelis, N., & Goldberg, A. E. (2020). Metaphorical language processing and amygdala activation in L1 and L2. Neuropsychologia.
The present study aims to investigate the neural correlates of processing conventional figurative language in non-native speakers in a comparison with native speakers. Italian proficient L2 learners of German and German native speakers read conventional metaphorical statements as well as literal paraphrases that were comparable on a range of psycholinguistic variables. Results confirm previous findings that native speakers show increased activity for metaphorical processing, and left amygdala activation increases with increasing Metaphoricity. At the whole-brain level, L2 learners showed the expected overall differences in activation when compared to native speakers (in the fronto-temporal network). But L2 speakers did not show any distinctive activation outside the caudate nucleus as Metaphoricity increased, suggesting that the L2 speakers were less affected by increasing Metaphoricity than native speakers were. With small volume correction, only a single peak in the amygdala reached threshold for L2 speakers as Metaphoricity increased. The findings are consistent with the view that metaphorical language is more engaging for native speakers but not necessarily for L2 speakers.
Floyd, S. & Goldberg, A. E. (2020). Children make use of relationships across meanings in word learning. Journal of Experimental Psychology: Learning, Memory and Cognition.
Many words are associated with more than a single meaning. Words are sometimes “ambiguous,” applying to unrelated meanings, but the majority of frequent words are “polysemous” in that they apply to multiple related meanings. In a preregistered design that included two tasks, we tested adults’ and 4½-7-year-old children’s ability to learn 4 novel polysemous words or 4 novel ambiguous words. Both children and adults demonstrated a polysemy over ambiguity learning advantage on each task after exposure, showing better learning of novel words with multiple related meanings than novel words with unrelated meanings. Stimuli in the polysemy condition were designed and then normed to guard against learners relying on a simple definition in order to distinguish the multiple target meanings for each word from foils. We retested available participants after a week-long delay without providing additional exposure and found that adults’ performance remained strong in the polysemy condition in one task, and children’s performance remained strong in the polysemy condition in both tasks. We conclude that participants are adept at learning polysemous words that vary along multiple dimensions. Current results are consistent with the idea that ambiguous meanings of a word compete, but polysemous meanings instead reinforce one another.
Ziegler, J., Bencini, G., Goldberg, A. E., & Snedeker, J. (2019). How abstract is syntax? Evidence from structural priming. Cognition.
In 1990, Bock and Loebell found that passives (e.g., The 747 was radioed by the airport’s control tower) can be primed by intransitive locatives (e.g., The 747 was landing by the airport’s control tower). This finding is often taken as strong evidence that structural priming occurs on the basis of a syntactic phrase structure that abstracts across lexical content, including prepositions, and is uninfluenced by the semantic roles of the arguments. However, all of the intransitive locative primes in Bock and Loebell contained the preposition by(by-locatives), just like the passive targets. Therefore, the locative-to-passive priming may have been due to the adjunct headed by by, rather than being a result of purely abstract syntax. The present experiment investigates this possibility. We find that passives and intransitive by-locatives are equivalent primes, but intransitive locatives with other prepositions (e.g., The 747 has landed near the airport control tower) do not prime passives. We conclude that a shared abstract, content-less tree structure is not sufficient for passive priming to occur. We then review the prior results that have been offered in favor of abstract tree priming, and note the range of evidence can be considerably narrowed and possibly eliminated, once effects of animacy, semantic roles, shared morphology, information structure, and rhythm are taken into account.
Tachihara, K. & Goldberg, A. E. (2019). Reduced competition effects and noisier representations in a second language. Language Learning.
Native speakers strongly disprefer novel formulations when a conventional alternative expresses the same intended message, presumably because the more conventional form competes with the novel form. In five studies, second language (L2) speakers were less influenced by competing alternatives than native speakers. L2 speakers accepted novel interpretable sentences more readily than native speakers, and were somewhat less likely to offer competing alternatives as paraphrases or to prefer competing alternatives in forced-choice tasks. They were unaffected by exposure to competing alternatives immediately before judgments. Reduced sensitivity to competing alternatives was confirmed by L2 speakers’ greater divergence from native speakers on judgments for novel formulations compared to familiar ones. Reduced sensitivity to competing alternatives also predicts noisier linguistic representations; consistent with this, L2 speakers performed worse on a verbatim recognition task, with performance correlating with more nativelike judgments. Proficiency was a modest predictor of judgments, but transfer effects were not.
Emberson, L., Loncar, N., Mazzei, C., Treves, I., & Goldberg, A. E. (2019). The blowfish effect: Children and adults use atypical exemplars to infer more narrow categories during word learning. Journal of Child Language.
Learners preferentially interpret novel nouns at the basic level (“dog”) rather than at a more narrow level (“Labrador”). This “basic-level bias” is mitigated by statistics: children and adults are more likely to interpret a novel noun at a more narrow label if they witness “a suspicious coincidence” — the word applied to 3 exemplars of the same narrow category. Independent work has found that exemplar typicality influences learners’ inferences and category learning. We bring these lines of work together to investigate whether the content (typicality) of a single exemplar affects the level of interpretation of words and whether an atypicality effect interacts with input statistics. Results demonstrate that both 4-5 year olds and adults tend to assign a narrower interpretation to a word if it is exemplified by an atypical category member. This atypicality effect is roughly as strong as, and independent of, the suspicious coincidence effect, which is replicated.
Tachihara, K., Pitcher, M., & Goldberg, A. E. (2019). Jessie and Gary or Gary and Jessie?: Feelings of closeness predict the order of proper name binomials in English and Japanese. Proceedings of the International Cognitive Science Society. Montreal, CA.
Notably, while English tends to prefer shorter before longer complements, Japanese displays the opposite tendency. Far less cross-linguistic work has investigated possible differences in the ordering of nouns within conjunctions (“binomials’), although one corpus study suggests that the same factors predict binomial ordering in Japanese and English. To investigate the latter issue experimentally, we report Japanese and English speakers’ productions of names of the members of couples that they knew personally, and found that both groups tended to name the member they felt closer to first, which we consider as an index of conceptual accessibility. Length (syllables/mora) was not a significant predictor in either language. Results confirm that conceptual accessibility is the most important factor in the ordering of binomial names in both languages. Differences in the preferred order of a verb’s complements must be attributed to other factors.
Floyd, S., Lew-Williams, C., & Goldberg, A. E. (2019). Children, more than adults, rely on similarity to access multiple meanings of words. Proceedings of the International Cognitive Science Society. Montreal, CA.
Past research has shown that adults can access multiple meanings for a label, but little work has examined how children process multiple meanings. We tested 48 4- to 7- year-old children and 48 adults in a touchscreen picture recognition task. Two meanings of the same word were displayed on successive trials, which varied according to whether the 2 meanings were unrelated (homonyms), related (polysemes), or repeated (same-meaning). Adults identified the second meaning more quickly than the first in all conditions and to the same extent. Children, however, identified the second meaning more quickly only on polysemy and same-meaning trials. This difference suggests that children are less capable of co-activating unrelated meanings, which raises the possibility that children must learn to do so over development. Despite the ubiquity of polysemy in language, our work is the first to show that children’s processing of word representations is organized by similarity.
Hernandez, A., Floyd, S., & Goldberg, A. E. (2019). Productivity depends on communicative intention and accessibility, not thresholds. Proceedings of the International Cognitive Science Society. Montreal, CA.
When do children extend a construction (“rule”) productively? A recent Threshold proposal claims that a construction is productive if and only if it has been witnessed applying to a sufficient proportion of cases and sufficiently few exceptions. An alternative proposal, Communicate + Access (C&A), argues that children extend a construction productively because they wish to express an intended message and are unable to access a “better” (appropriate and more conventional) way to do it. Accessibility, in turn, is negatively affected by interference from competing alternatives. In a (preregistered) experiment, 32 4-6-year-old children were provided with exposure to 2 mini-artificial languages for which the two proposals make opposite predictions. Results support the C&A proposal: children were more productive after witnessing 3 rule-following cases than after 5, due to differences in interference. We conclude that productivity is encouraged by a desire to communicate a message and is constrained by accessibility/interference.
Floyd, S., Goldberg, A. E., & Lew-Williams, C. (2019). Toddlers recognize multiple polysemous meanings and use them to infer additional meanings. Proceedings of the International Cognitive Science Society. Montreal, CA.
Up to 80% of words have multiple, related meanings (polysemy), yet work on early word learning has almost uniformly assumed one-to-one mappings between form and meaning. Using a looking-while-listening procedure, we present the first evidence that toddlers (n=32) can recognize multiple meanings for common nouns, e.g. collar of a dog, shirt collar. In an English-meaning condition, toddlers were tested on their ability to recognize multiple English meanings for polysemous words such as cap(e.g. a baseball cap and a bottle cap). Another condition prompted them with the same English words (e.g., cap), but target referents instead corresponded to the word’s polysemous extension in an unfamiliar language, (e.g., “lid” is a meaning for Spanish’s “cap”, tapa). Toddlers looked to the correct targets above chance in both trial types, but with greater accuracy on English-meaning trials, demonstrating a recognition of familiar word-meaning pairs andan ability to infer potential new meanings (Spanish-meaning trials).
Tachihara, K., Norman, K., Turk-Browne, N., & Goldberg, A. E. (2019). A generalization becomes suppressed in the context of exceptions. Proceedings of the International Cognitive Science Society. Montreal, CA.
There has been a great deal of interest in how generalizations and exceptions are represented and processed, but scant attention has been paid to the following question: do exceptional cases affect generalizations during learning? Here we suggest that they do. Novel words were paired with one of two preceding “classifiers.” Each <classifier+word> was assigned a unique image. Most words for one classifier were paired with images from a generalization semantic category, whereas a subclass of exception words were paired with images from a second category. All words for the second classifier appeared with a third baseline category. After initial exposure, participants used a computer mouse to choose one of two images for each <classifier+word>, in a design repeated over 8 blocks. With deviation over time toward the lure as the dependent measure, results showed that competition led to suppression of a generalization in the context of exceptions, suggesting that the generalization itself was affected by learned exceptions.
Explain me this: Creativity, Competition, and the Partial Productivity of Constructions. Adele E. Goldberg (2019, Princeton University Press). Table of Contents.
Goldberg, A. E. (2018). The sufficiency principle hyperinflates the cost of productivity. Linguistic Approaches to Bilingualism, 8(6), 727-732.
Whenever the number of exceptions to a rule reaches its maximum, as it often does in the examples cited, the SP claims that learners must witness and retain all other cases that potentially follow a rule actually following the rule, in order for the rule to become “productive.” While children have been argued to be conservative learners, the SP takes conservatism to a whole new level.
Barak, L., Floyd, S., & Goldberg, A. E. (2019) Modeling the acquisition of words with multiple meanings. Proceedings of the Society for Computation in Linguistics (SCiL).
Learning vocabulary is essential to successful communication. Complicating this task is the underappreciated fact that most common words are associated with multiple senses (are polysemous ) (e.g., baseball cap vs. cap of a bottle), while other words are homonymous, evoking meanings that are unrelated to one another (e.g., baseball bat vs. flying bat ). Models of human word learning have thus far failed to represent this level of naturalistic complexity. We extend a feature-based computational model to allow for multiple meanings, while capturing the gradient distinction between polysemy and homonymy by using structured sets of features. Results confirm that the present model correlates better with human data on novel word learning tasks than the existing feature-based model.
Schwab, J., Lew-Williams, C., & Goldberg, A. E. (2018). When regularization gets it wrong: Children over-simplify language input only in production. Journal of Child Language.
Children tend to regularize their productions when exposed to artificial languages, an advantageous response to unpredictable variation. But generalizations in natural languages are typically conditioned by factors that children ultimately learn. In two experiments, adult and six-year-old learners witnessed two novel classifiers, probabilistically conditioned by semantics. Whereas adults displayed high accuracy in their productions—applying the semantic criteria to familiar and novel items—children were oblivious to the semantic conditioning. Instead, children regularized their productions, over-relying on only one classifier. However, in a two-alternative forced-choice task, children’s performance revealed greater respect for the system’s complexity: they selected both classifiers equally, without bias toward one or the other, and displayed better accuracy on familiar items. Given that natural languages are conditioned by multiple factors that children successfully learn, we suggest that their tendency to simplify in production stems from retrieval difficulty when a complex system has not yet been fully learned.
Tachihara, K. & Goldberg, A. E. (2018) L2 speakers are more accepting of unconventional language than native speakers. CUNY.
Native speakers strongly disprefer novel or unconventional combinations of verbs in argument structure constructions when a conventional competing alternative (“CA”) already exists (e.g., He forced that she compete; CA: He forced her to compete). L2 speakers also rate unconventional sentences as less acceptable than familiar formulations, but they are somewhat more tolerant than L1 speakers are (Experiment 1). In a paraphrase task, L2 speakers were less likely than L1 speakers to provide CAs as paraphrases of the unconventional sentences (Experiment 2), and in a 2-alternative-forced-choice task, L2 speakers were somewhat less likely than L1 speakers to prefer the CAs over the unconventional formulations (Experiment 3). Even when L2 speakers did prefer the CAs in the 2AFC task, they were more generous with their ratings of the unconventional alternatives than L1 speakers were (Exp. 3). We investigated whether inducing a greater awareness of relevant CAs would lead L2’s judgments on unconventional sentences to align more closely with native speakers,’ but found no effect (Experiment 4). A final verbatim memory recognition test showed lower performance for L2 speakers, with performance correlating with judgements (Experiment 5). A final meta-analysis of judgment data across studies finds proficiency to be a modest predictor of judgments, while transfer effects are not predictive in the present studies in any straightforward way.
Perek, F. & Goldberg, A. E. (2017). Linguistic generalization on the basis of function and constraints on the basis of statistical preemption. Cognition.
How do people learn to use language in creative but constrained ways? Experiment 1 investigates linguistic creativity by exposing adult participants to two novel word order constructions that differ in terms of their semantics: One construction exclusively describes actions that have a strong effect; the other construction describes actions with a weaker but otherwise similar effect. One group of participants witnessed novel verbs only appearing in one construction or the other, while another group witnessed a minority of verbs alternating between constructions. Subsequent production and judgment results demonstrate that participants in both conditions extended and accepted verbs in whichever construction best described the intended message. Unlike related previous work, this finding is not naturally attributable to prior knowledge of the likely division of labor between verbs and constructions. In order to investigate how speakers learn to constrain generalizations, Experiment 2 includes one verb (out of 6) that was witnessed in a single construction to describe both strong and weak effects, essentially statistically preempting the use of the other construction. In this case, participants were much more lexically conservative with this verb and other verbs, while they nonetheless displayed an appreciation of the distinct semantics of the constructions with new novel verbs. Results indicate that the need to better express an intended message encourages generalization, while statistical preemption constrains generalization by providing evidence that verbs are restricted in their distribution.
Barak, L. & Goldberg, A. E. (2017). Modeling the Partial Productivity of Constructions. American Association for Artificial Intelligence Spring Symposium (AAAI).
People regularly produce novel sentences that sound native-like (e.g., she googled us the information), while they also recognize that other novel sentences sound odd, even though they are interpretable (e.g., ? She explained us the information). This work offers a Bayesian, incremental model that learns clusters that correspond to grammatical constructions of different type and token frequencies. Without specifying in advance the number of constructions, their semantic contributions, nor whether any two constructions compete with one another, the model successfully generalizes when appropriate while identifying and suggesting an alternative when faced with overgeneralization errors. Results are consistent with recent psycholinguistic work that demonstrates that the existence of competing alternatives and the frequencies of those alternatives play a key role in the partial productivity of grammatical constructions. The model also goes beyond the psycholinguistic work in that it investigates a role for constructions’ overall frequency.
Comparing Computational Cognitive Models of Generalization in a Language Acquisition Task. Libby Barak, Adele E. Goldberg and Suzanne Stevenson. EMNLP 2016. [pdf]
Natural language acquisition relies on appropriate generalization: the ability to produce novel sentences, while learning to restrict productions to acceptable forms in the language. Psycholinguists have proposed various properties that might play a role in guiding appropriate generalizations, looking at learning of verb alternations as a testbed. Several computational cognitive models have explored aspects of this phenomenon, but their results are hard to compare given the high variability in the linguistic properties represented in their input. In this paper, we directly compare two recent approaches, a Bayesian model and a connectionist model, in their ability to replicate human judgments of appropriate generalizations. We find that the Bayesian model more accurately mimics the judgments due to its richer learning mechanism that can exploit distributional properties of the input in a manner consistent with human behaviour.
Subtle Implicit Language Facts Emerge from the Functions of Constructions. Adele E. Goldberg. 2016. Frontiers in Psychology 6. doi: 10.3389/fpsyg.2015.02019 [web-link]
Much has been written about the unlikelihood of innate, syntax-specific, universal knowledge of language (Universal Grammar) on the grounds that it is biologically implausible, unresponsive to cross-linguistic facts, theoretically inelegant, and implausible and unnecessary from the perspective of language acquisition. While relevant, much of this discussion fails to address the sorts of facts that generative linguists often take as evidence in favor of the Universal Grammar Hypothesis: subtle, intricate, knowledge about language that speakers implicitly know without being taught. This paper revisits a few often-cited such cases and argues that, although the facts are sometimes even more complex and subtle than is generally appreciated, appeals to Universal Grammar fail to explain the phenomena. Instead, such facts are strongly motivated by the functions of the constructions involved. The following specific cases are discussed: (a) the distribution and interpretation of anaphoric one , (b) constraints on long-distance dependencies, (c) subject-auxiliary inversion, and (d) cross-linguistic linking generalizations between semantics and syntax.
Ellipsis by constructions. Adele E. Goldberg and Florent Perek. In Handbook of Ellipsis. Edited by Jeroen van Craenenbroeck & Tanja Temmerman. Oxford University Press:
The existence of elliptical constructions in languages is motivated by the Gricean preference to avoid saying more than is necessary. We suggest interpretation is recovered by a semantic pointer function that is independently motivated and consistent with psycholinguistic evidence. This article reviews evidence that ellipsis in language is best understood as licensed by particular constructions, each with its own form and functional properties. Our account of gapping is quite general in that it includes cases of traditional “argument cluster coordination”; it is at the same time more restrictive than other accounts in including a constraint on register. We compare gapping and other ellipsis constructions in French and English with an emphasis on their differences. Finally, we discuss derivational approaches to ellipsis, and concluded that a constructionist approach is more promising than a single, over-arching rule-based approach since it is in a better position to capture distinctions as well as commonalities among ellipsis constructions, within and across languages.
Conventional metaphorical sentences such as She’s a sweet child have been found to elicit greater amygdala activation than matched literal sentences (e.g., She’s a kind child). In the present fMRI study, this finding is strengthened and extended with naturalistic stimuli involving longer passages and a range of conventional metaphors. In particular, a greater number of activation peaks (four) were found in the bilateral amygdala when passages containing conventional metaphors were read than when their matched literal versions were read (a single peak); while the direct contrast between metaphorical and literal passages did not show significant amygdala activation, a parametric analysis revealed that BOLD signal changes in the left amygdala correlated with an increase in metaphoricity ratings across all stories. Moreover, while a measure of complexity was positively correlated with increase in activation of a broad bilateral network mainly involving the temporal lobes, complexity was not predictive of amygdala activity. Thus, the results suggest that amygdala activation is not simply a result of stronger overall activity related to language comprehension, but is more specific to the processing of metaphorical language.
Neural systems involved in processing novel linguistic constructions and their visual referents. Matthew A. Johnson, Nick Turk-Browne, and Adele E. Goldberg. 2015. Language, Cognition, and Neuroscience:
In language, abstract phrasal patterns provide an important source of meaning, but little is known about whether or how such constructions are used to predict upcoming visual scenes. Findings from two fMRI studies indicate that initial exposure to a novel construction allows its semantics to be used for such predictions. Specifically, greater activity in the ventral striatum, a region sensitive to prediction errors, was linked to worse overall comprehension of a novel construction. Moreover, activity in occipital cortex was attenuated when a visual event could be inferred from a learned construction, which may reflect predictive coding of the event. These effects disappeared when predictions were unlikely: that is, when phrases provided no additional information about visual events. These findings support the idea that learners create and evaluate predictions about new instances during comprehension of novel linguistic constructions.
Judgment evidence for statistical preemption: It is relatively better to vanish than to disappear a rabbit, but a lifeguard can equally well backstroke or swim children to shore. Clarice Robenalt & Adele E. Goldberg. 2015. Cognitive Linguistics. 26.3: 467-504.[pdf]
How do speakers know when they can use language creatively and when they cannot? Prior research indicates that higher frequency verbs are more resistant to overgeneralization than lower frequency verbs with similar meaning and argument structure constraints. This result has been interpreted as evidence for conservatism via entrenchment, which proposes that people prefer to use verbs in ways they have heard before, with the strength of dispreference for novel uses increasing with overall verb frequency. This paper investigates whether verb frequency is actually always relevant in judging the acceptability of novel sentences or whether it only matters when there is a readily available alternative way to express the intended message with the chosen verb, as is predicted by statistical preemption. Two experiments are reported in which participants rated novel uses of high and low frequency verbs in argument structure constructions in which those verbs do not normally appear. Separate norming studies were used to divide the sentences into those with and without an agreed-upon preferred alternative phrasing which would compete with the novel use for acceptability. Experiment 2 controls for construction type: all target stimuli are instances of the caused-motion construction. In both experiments, we replicate the stronger dispreference for a novel use with a high frequency verb relative to its lower frequency counterpart, but only for those sentences for which there exists a competing alternative phrasing. When there is no consensus about a preferred way to phrase a sentence, verb frequency is not a predictive factor in sentences’ ratings. We interpret this to mean that while speakers prefer familiar formulations to novel ones, they are willing to extend verbs creatively if there is no readily available alternative way to express the intended meaning.
L2 learners do not take competing alternative expressions into account the way L1 learners do. Clarice Robenalt & Adele E. Goldberg. 2015. Language Learning: [pdf]
The present study replicates the findings in Robenalt & Goldberg (to appear, CogLing–immediately above) with a group of native speakers and critically extends the paradigm to non-native speakers. Recent findings in second language acquisition suggest that second language (L2) learners are less able to generate online expectations during language processing, which in turn predicts a reduced ability to differentiate between novel sentences that have a competing alternative and those that do not. We test this prediction and confirm that while L2 speakers display evidence of learning from positive exemplars, they show no evidence of taking competing grammatical alternatives into account, except at the highest quartile of speaking proficiency in which case L2 judgments align with native speakers.
A-adjectives, statistical preemption, and the evidence: Reply to Yang (2015). Adele E. Goldberg and Jeremy K. Boyd. 2015. Language 91 11: 184-197. pdf
A certain class of English adjectives known as a-adjectives resists appearing attributively as prenominal modifiers (e.g., ??the afraid boy, ??the asleep man). Boyd & Goldberg (2011) had offered experimental evidence suggesting that the dispreference is learnable on the basis of categorization and statistical preemption: repeatedly witnessing predicative formulations in contexts in which the attributive form would otherwise be appropriate. The present paper addresses Yang (2015)’s counterproposal for how a-adjectives are learned, and his instructive critique of statistical preemption. The counterproposal is that children receive evidence that a-adjectives behave like locative particles in occurring with certain adverbs such as far and right. However, in an analysis of the 450 million word COCA corpus, the suggested adverbial evidence is virtually non-existent (e.g., *far alive; *straight afraid). In fact, these adverbs occur much more frequently with typical adjectives (e.g., far greater, straight alphabetical). Furthermore, relating a-adjectives to locative particles does not provide evidence of the restriction, because locative particles themselves can appear as prenominal modifiers (the down payment, the outside world). The critique of statistical preemption is based on a 4.3 million word corpus analysis of child directed speech that suggests that children cannot amass the requisite evidence before they are three years old. While we clarify which sorts of data are relevant to statistical preemption, we concur that the required data is relatively sparsely represented in the input. In fact, recent evidence suggests that children are not actually cognizant of the restriction until they are roughly ten years old, an indication that input of an order of magnitude more than 4.3 million words may be required. We conclude that a combination of categorization and statistical preemption is consistent with the available evidence of how the restriction on a-adjectives is learned.
One among many: anaphoric one and its relationship to numeral one. Adele E. Goldberg & Laura A. Michaelis. 2016. Cognitive Science: [pdf]
One anaphora (e.g., She has a better one) has been used as a key diagnostic in syntactic analyses of the English noun phrase, and ‘one-replacement’ has also figured prominently in debates about the learnability of language. However, much of this work has been based on faulty premises, as a few perceptive researchers, including Ray Jackendoff, have made clear. Abandoning the view of anaphoric one (a-one) as a form of syntactic replacement allows us to take a fresh look at various uses of the word one. In the present work, we investigate its use as a cardinal number (1-one) in order to better understand its anaphoric use. Like all cardinal numbers, 1-one can only quantify an individuated entity and provides an indefinite reading by default. Owing to unique combinatoric properties, cardinal numbers defy consistent classification as determiners, quantifiers, adjectives or nouns. Once the semantics and distribution of cardinal numbers including 1-one are appreciated, many properties of a-one follow with minimal stipulation. We claim that 1-one and a-one are distinct but very closely related lexemes. When 1-one appears without a noun (e.g., Take one), it is nearly indistinguishable from a-one (e.g., take one)—the only differences being interpretive (1-one foregrounds its cardinality while a-one does not) and prosodic (presence versus absence of primary accent). While we ultimately argue that a family of constructions is required to describe the full range of syntactic contexts in which one appears, the proposed network accounts for properties of a-one by allowing it to share (inherit) most of its syntactic and interpretive constraints from its historical predecessor, 1-one.
Generalizing beyond the input: the functions of the constructions matter: Florent Perek & Adele Goldberg. 2015. Journal of Memory and Language 84: 109-127.
A growing emphasis on statistics in language learning raises the question of whether learning a language consists wholly in extracting statistical regularities from the input. In this paper we explore the hypothesis that the functions of learned constructions can lead learners to use language in ways that go beyond the statistical regularities that have been witnessed. The present work exposes adults to two novel word order constructions that differed in terms of their functions: one construction but not the other was exclusively used with pronoun undergoers. In Experiment 1, participants in a lexicalist condition witnessed three novel verbs used exclusively in one construction and three exclusively in the other construction; a distinct group, the alternating condition, witnessed two verbs occurring in both constructions and two other verbs in each of the constructions exclusively. Production and judgment results demonstrate that participants in the alternating condition accepted all verbs in whichever construction was more appropriate, even though they had seen just two out of six verbs alternating. The lexicalist group was somewhat less productive, but even they displayed a tendency to extend verbs to new uses. Thus participants tended to generalize the constructions for use in appropriate discourse contexts, ignoring evidence of verb-specific behavior, especially when even a minority of verbs were witnessed alternating. A second experiment demonstrated that participants’ behavior was not likely due to an inability to learn which verbs had occurred in which constructions. Our results suggest that construction learning involves an interaction of witnessed usage together with the functions of the constructions involved.
Tuning in to the verb-particle construction in English. Adele E. Goldberg. to appear. In Léa Nash and Pollet Samvelian (eds.) Approaches to Complex Predicates:
This work investigates English verb particle combinations (e.g., put on) and argues that item-specific and general information are needed and should be related within a default inheritance hierarchy. When verb particle combinations appear within verb phrases, a tripartite phrasal syntax is defended, whether or not the V and P are adjacent (e.g., She put on the wrong shoes; she put the wrong shoes on). The < V NP P > order is motivated as the default word order by explicitly relating a verb-particle construction to the caused-motion construction (e.g., she put the shoes on her feet). Well-known and independently needed processing considerations related to complement length, information status, and semantics motivate system-wide generalizations that can serve to override the default word order. Lexical verb-particle combinations (e.g., a pickup truck; a showdown) and an idiomatic case, V-off are also briefly discussed as providing further evidence for the need for both item-specific and more general constructions.
Compositionality. Adele E. Goldberg. 2016. In N. Reimer (ed.) Routledge Handbook of Semantics. 419-430.
How do people glean meaning from language? A Principle of Compositionality is generally understood to entail that the meaning of every expression in a language must be a function of the meaning of its immediate constituents and the syntactic rule used to combine them. This paper explores perspectives that range from acceptance of the principle as a truism, to rejection of the principle as false. Controversy has arisen basedon the role of extra-constituent linguistic meaning (idioms; certain cases of paradigmatic morphology; constructional meanings; intonation), and context (e.g., metonymy; the resolution of ambiguity and vagueness).